Plenary Session: Measuring Consumer Confusion in Trademark
Infringement
Facilitator: Barton Beebe (NYU)
Lanham Act: confusion is vaguely defined. Used to include “purchasers” but Congress
deleted that phrase. 43(a)’s language is
even broader. Flexible and slippery.
Panelists: Joel Steckel (NYU Stern School of Business)
Topic: Consumer Confusion Surveys Used in Litigation, Commenting
on: Robert H. Thornburg, Trademark
Surveys: Development of Computer-Based Survey Methods, 4 J. Marshall Rev.
Intell. Prop. L. 91 (2005)
Wants to talk about dilution. Thornburg paper: not a very good paper, but
that’s because it does two things—a catalog of different survey types/principles,
then talks about difficulty of getting internet surveys into evidence. Much of
what he said in 2004 is obsolete in 2014, when internet surveys are by and
large admitted with some exceptions.
Drilling down into one aspect: dilution surveys. If confusion is
slippery, dilution makes that look like sandpaper. Harm the reputation/impair
the distinctiveness of a mark: what does that mean?
Evident from usual surveys that courts accept that courts
don’t really know what they’re trying to measure. Exxon survey: Nike v. Nikepal, where
respondents were simply asked what if anything came to your mind when I first
said the word Nikepal, 79% said Nike.
Makes you wonder, what did the other 21% say? That is considered evidence; it’s evidence of
association, but doesn’t correspond to impaired distinctiveness or harmed
reputation.
Associative network memory model of brands. Knowledge is a network of bits of info: TM or
associated logo, slogans, etc., and associations—anything that comes to mind
when you think of a brand name. When the word Coke is presented to customers,
they think of “tradition,” “nostalgia,” etc. and Pepsi is more
excitement/use/taste. Dilution offense
occurs when a junior brand enters and has another set of brand associations. Tarnishment if a new association is negative
and somehow gets attached to the senior brand: if Victor’s Little Secret causes
consumers to associate Victoria’s Secret with pornography. Test it by seeing if test group associates VS
with porn. Blurring: distinctiveness would
be impaired if link between Nike and sports was weakened; reaction time to
associating Nike with sports would be a measure of blurring.
Takeaways: dilution has been difficult to measure because it
wasn’t defined well. He argues that his
definitions are suitable theoretically based definitions. Marketer’s perspective: a sensible definition
would include the potentially dilutive power of mixing and adding brand
associations. Measuring the degree to which brand asociations are held and how
easily they are recalled can provide measures consistent w/the language of the
law.
Lisa Larrimore Ouellette (Stanford)
Topic: Cognitive/Psychological Approaches to Modeling When
and How Consumers Get Confused, Commenting on: Thomas R. Lee et al., An
Empirical and Consumer Psychology Analysis of Trademark Distinctiveness, 41
Ariz. St. L.J. 1033 (2009); Thomas R. Lee et al., Trademarks,
Consumer Psychology, and the Sophisticated Consumer, 57 Emory L.J. 575
(2008)
There is a fair amount of disagreement about TM’s goals and
therefore about what facts matter. Take
courts at their word: consumer mindset/confusion is what we care about. Multifactor test for likely confusion. Buyer sophistication: there’s a whole
literature about consumer care in buying choices and courts aren’t paying
attention to that, which we should bring in. No clear direction for likely
confusion; another paper looks at likelihood of bridging the gap.
Sophisticated consumer paper: courts have ID’d many factors
as relevant to this: low price of goods; purchase complexity, frequency of
purchase; education, age, gender, and income of buyers; professional buyers or
hobbyists. But little attention to consumer psych literature. That literature discusses consumer motivation
as well as consumer ability to exert effort to make distinctions; they need both. Motivation can depend on the realm and on the
person (some people have greater needs for cognition than others). Longer time to make decision without
distraction can make difference. Courts
make generalizations about low price, but that’s not the relevant
question. Low financial risk may be high
physical or social risk, creating more motivation.
Consumers who saw Mercedes Benz computers were more likely
to think that Cadillac would bridge the gap to consumers too. Those test subjects with more experience
buying consumers or more education were more likely to be confused—cuts against
judicial doctrine.
Secondary meaning survey: created fictitious marks with no
secondary meaning and mimicked typical trademark use on products—Chocolate
Abundance, Party Hat, Fudge Covered Cookies etc. for chocolate covered
cookies. If you believe the Abercrombie
spectrum means something you’d expect some linear progression. What they found was that consumers were
unlikely to see generic terms as indicating source, but for the rest it was all
the same. And even generic term, over ¼
of consumers saw it as indicating a brand name.
Presentation matters: large font/ordinary TM place makes TM perception
more likely, small font/placement makes TM perception less likely.
Abercrombie may
serve other interests, like protecting competitors—can still be reasons to
encourage parties to select more arbitrary/fanciful terms over descriptive
ones. Still, it’s not good for telling
us consumer mindset, and we need other measures of distinctiveness and
strength. Inherent distinctiveness is a
very weak proxy for what consumers are looking at.
Could test other factors like similarity of marks and
proximity of goods in similar ways.
Courts’ general assumptions may be unreliable.
Mark McKenna (Notre Dame)
Topic: Qualitative Studies of Consumers Becoming Confused
During Shopping, Commenting on: G. Miaoulis & N. D’Amato, Consumer
Confusion and Trademark Infringement, J. Marketing 48-55 (April 1978); Vincent-Wayne
Mitchell & Vassilios Papavassiliou, Exploring
Consumer Confusion in the Watch Market, 15 Marketing Intelligence &
Planning 164-172 (1997)
There aren’t very many good studies about confusion. One reason: marketing folks aren’t interested
in confusion as defined in TM law. They’re intensely interested in how
marketplace practices affect a brand, but there’s a very important distinction
between brands and marks; brands are much broader concepts including associated
meaning which can be affected by a much wider range of practices that might or
might not involve TMs/confusion.
Marketers are much more focused on the question of harm. Those might be in
cases where consumers know accurately that a new product comes from the
existing source (might be thought to simulate 100% confusion)—is there any harm
to the brand under those circumstances?
TM law accepts claims of harm much more readily than empirical evidence
justifies. Studies are also focused on
harm to the brand, not harm to consumers except maybe incidentally. So you get
statements like the M&D’A study—confusion in many circumstances won’t harm
consumers because consumers don’t care. Thus these studies press normatively to
include in the definition of confusion a broader range of effects than TM
academics care; trying to shape legal doctrine to test the things they care
about.
First study: attempts to measure extent to which consumers
exposed to new product that shares features—here, packaging of mints, features
of the brand but not the TM—will generalize characteristics from the known
product to the new product. Authors
count stimulus generalization as confusion! They want this to count even if
normal purchasing conditions make other characteristics distinguish the
products, which they call “intellectually discriminating” between the
products. “Confusion” despite the
absence of source confusion on the theory that this generalization harms the
brand by making it less “differentiated.” And indeed they so find—people think
the new product will be refreshing and minty or will taste good. That, they
think, is terrible! Consumers porting
info from Tic-Tac to other mints makes Tic-Tac no longer as distinctive.
tic-tacs, Dynamints, and Mighty Mints |
Conventional TM perspective: completely irrelevant to doctrinal convention. TM is concerned with one kind of generalization: source information—who is responsible b/c of the presence of mark. Non-source info shouldn’t matter; and porting info for reasons unrelated to protectable to mark should also be irrelevant. Normatively this is correct; ability to generalize is often unqualifiedly good for consumers and for competition even if bad for particular brand owner. Effective communication of “my product is minty and tastes good too” is called competition. TM’s job is not to prevent people from selling products with similar desirable characteristics. That’s why we approve comparative advertising. How we describe a new restaurant: almost always in relation to something you already know.
Modern TM law becomes more intelligible if you read it
through the lens of these studies. Doctrines that most befuddle academics are
b/c courts squeeze into TM the broader concerns in these studies—dilution,
initial interest confusion, post sale confusion.
Mitchell & Papavassiliou study: followed customers
around watch store and asked them what aspects made the experience more
challenging and what they did to overcome those challenges. Consumers say: fragmented nature of market;
newness of tech in watches; hidden nature of watch movement; role fashion; too
many brands/shops; low frequency of purchase; purchase of watches as gifts. The
word for these effects isn’t confusion, but contextual factors that interfere
with decisions in shopping/raise search costs, but not confusion. But maybe
that means confusion isn’t well defined.
Very few of these interferences are things TM has anything
to say about. Maybe that’s a good reason to avoid thinking of TM as a means to
combat search costs but rather a particular sort of interference.
Info overload/overchoice.
Consumers can struggle to choose when there’s too much info. Study tries
to figure out consumer strategies—clarifying goals, narrowing choice, seeking
additional info. This 1998 study could usefully be updated for new contexts
like the internet. More interesting:
what relationship TM has to the problem of overchoice. Paper simply asserts that counterfeiting is a
source of confusion and law can fight that, but the study is from a store
apparently lacking in counterfeits. Paper says subbranding is contributor to
confusion, undertaken by brand owners themselves as a way of differentiating
products: abundance of functionally similar products differentiated on a
fashion platform. We think of TM as reducing search costs by making search less
time consuming, but TM also affects differentiation in the market, and has a
lot to say how close products may be and how they must be
differentiated—product configuration like color of band.
To what extent is TM contributing to and reducing search
costs, and how do those net out? If TM incentivizes lots of dimensional
differentiation, that can increase search costs.
Studies differ but have common pressure to define confusion
more broadly. Search costs rhetoric has made it easier to redefine confusion in
this way. As Mitchell study illuminates,
there are lots of other kinds of search costs, and it’s too easy to call those
confusion. Academics: if you accept that
TM doctrine has responded to these papers’ concerns even if the doctrine isn’t
well suited, then we have to do more than point out the irrelevance of these
types of confusion. Vacuum about types of confusion that ought to be relevant;
studies focus on irrelevant question. Where empirical evidence is, doctrine
moves. To resist, need to do some good empirical work about when consumers are confused in the way we want to mean
it.
I’ve been asked to
speak about the big data approach to consumer confusion. One promise, likely
illusory, of big data is that we need not care any more about causality; we can
merely find correlations, however unexpected, and react to them. But trademark infringement, and even
dilution, requires a causal narrative as currently understood: this use created
this mental state in the consumers who saw it.
As such, big data may not be our savior.
Instead, large datasets may be able to change the theory of what trademark
infringement (and even dilution) is, the same way that previous advances in
marketing and consumer psychology led some to reconceptualize what trademark
infringement and dilution are, but that will be a normative choice even more
than an empirical one.
We’ve seen one
American case really take a stab at using “big data,” in the form of Google ad
clickthrough rates. That’s 1-800
Contacts, Inc. v. Lens.com, Inc., decided by the 10th Circuit in
2013. The court says that the theory of
IIC is that consumers seeking 1-800-Contacts (which we know, the court says,
because they searched for the term) clicked on a Lens.com ad while believing it
was a 1-800 site and, though no longer confused when they arrived at a Lens.com
site, were nonetheless diverted. We have
no idea how many were confused when they clicked and how many were not confused
but rather were seeking a possible alternative to 1-800, but the court says
that we do know the upper bound of the former number: the total number of clickthroughs,
which was a tiny fraction of the impressions.
For ads without the 1-800 Contacts trademark in the ad copy, Lens.com
got a 1.5% clickthrough rate. This was
too tiny to count as likely confusion even if all clickthroughs were the result
of IIC, so Lens.com couldn’t be liable for infringement based on those
ads. But ad clickthrough rates are always very low, so even a straight-up
counterfeiter as in Rosetta Stone v. Google would seem to escape liability for
infringement if you measure confusion in this manner.
This brings me to an
argument made in a 2000 article by Alex Simonson, Survey Design in False
Advertising Cases: he argues that we need to pay attention to attention versus
comprehension. We might have a survey
where a small percentage of people notice a particular product feature, say a
picture of a black rooster and the label Gallo Nero on the neck seal of a
bottle of wine, but those who do almost all make a connection to E&J Gallo
wines. Or our survey might show that
many people notice the neck seal, but only a few make the connection to E&J
Gallo. Simonson argues that those are
different results even if they produce the same net number of so called
confused consumers: for legal purposes, we ought to be concerned more about
comprehension than attention. Attention,
after all, varies in the real world, and so artificial attempts to measure it
may not be very good—but if we look at the percentage of people within the set
who noticed a feature or symbol who were confused by that feature or symbol, we
have a result that may be more generalizable as well as more important.
I think this point
applies beyond surveys: it’s a truism that most advertising is completely
useless—and now we have eye tracking studies showing that most people literally
do not see most ads, to which one English court has referred in its reasoning,
and MRIs showing that a lot of the time our brains don’t even react to ads, not
even bothering to process them. In that
sense, there is no such thing as an ad that is likely to confuse, because it’s not likely to reach us in the first
place. But the people who don’t notice
the ad are arguably irrelevant to any confusion inquiry. It’s only those who notice who might be confused. So, while I hate initial interest
confusion and think Lens.com is right
on the merits, I don’t think the analysis is right. All we know is the number of people who
clicked on the link—but we don’t know how big a percentage that is of the
number who noticed the link. It might be
that almost everybody who looked at that link clicked on it, possibly because
they were confused, and that scenario ought to concern us if it occurs,
especially to the extent that there are no countervailing benefits from having
the ad—and it’s hard to explain how the ad might benefit consumers who don’t
notice it at all, though one might possibly construct an argument about
chilling effects on truthful advertisers.
Also, we need a
concept of relevant confusion, or materiality.
Something that people mostly don’t pay attention to because they don’t
care about it shouldn’t be deemed confusing.
The broader takeaway is that we still need a theory to explain what we
should care about and why; big data do not remove the need for big ideas.
That leads me into
the papers on which the organizers specifically asked me to comment: Stefan
Bechtold & Catherine Tucker, Trademarks,
Triggers and Online Search, J. Empirical Leg. Stud. (forthcoming), and Lisa
Larrimore Ouellette’s The Google
Shortcut to Trademark Law, 102 Calif. L. Rev. 351 (2014).
Bechtold and Tucker
used a large dataset to explore the effects of Google’s policy changes allowing
more competitor keyword purchases of trademarks in Germany and France. Basically, they found little net change in
the number of consumers who ended up on the trademark owner’s website, but a
change in composition. They divided
searches into navigational searches, where the consumer is searching for the
keyword because she is directly interested in using the search engine as a
shortcut to find a specific webpage such as the trademark owner’s website, and
non-navigational searches where the consumer is doing something else. They estimated which were which by looking at
how long the search was—so iPhone would be navigational but “how to restart my
iPhone would be non-navigational”—and some other contextual factors.
After Google’s
liberalization, navigational searches became less likely to lead to the
trademark owner’s website, while non-navigational searches were more likely to
do so. Percentagewise, they classified
20% of searches as purely navigational, while 80% were non-navigational. The
policy change was associated in a 9.2 % decrease in consumers visiting the
trademark owners’ websites when they used a search phrase that exactly matched
the trademark. But consumers who were searching using the trademark alongside
other words were more likely to reach the trademark owners’ websites in 14.7%
of cases.
What can we glean
from this? The authors rightly recognize
the limits of their results. While they
suggest that a navigational searcher’s search process might be “impeded” by unauthorized
use of a mark, “as her attention is drawn to many third-party websites in which
the searcher is not interested,” we don’t know whether she perceives any
impediment. A searcher committed to
finding the trademark owner’s own site can usually determine without clicking
which site is which, especially with the prominent brands the authors
tested. David
Franklyn and David Hyman have shown that consumers are often confused about
whether a search result is organic or paid, but rarely confused about the
underlying source of those ads.
What the authors can test—and thus what their data might
push trademark theory to care about—is whether trademark owners suffer any loss
of consumer visits from the policy change. They conclude that their findings do
provide an upper bound to potential negative effects of the policy change on
trademark owners: they only suffer from negative consequences within the
subclass of navigational users, only 20% of searches. In European law, arguably
these negative effects have something to do with the “investment function” and
the “advertisement function” of trademarks, though I have to admit I don’t
really know what those are other than ways of stating ownership claims
regardless of any effect on consumers.
Even if the consumer began by wanting to visit the trademark owner’s
site, we can’t say without knowing more that she’s worse off under Google’s
liberalized policy. Maybe the new
choices made her rethink her initial desires.
If the ads weren’t confusing, we need some other reason to say that’s
wrongful, and even the European approach doesn’t explain why diversion is a
wrong to the consumer.
I’m intrigued by
some subsidiary findings for the light they can throw on what we don’t know about consumers: First,
“searches on Google appear to be consistently associated with fewer visits to
the trademark owner’s site and more searches and activity before a visit to the
trademark owner’s site even before the policy change” compared to Bing or
Yahoo! Second, “relative to France,
searches originating in Germany are less likely to lead to a trademark owner’s
site and also … [German] searchers are more likely to engage in multiple
searches prior to a visit to a trademark owner’s site.” National origin and Google brand loyalty
appear to be independent effects on the extent to which users are precommitted
to trademarks, which ought to shake our confidence that we know what matters to
consumers, much less why.
More generally: the
paper provides important information about consumer behavior, but not about
what consumers are thinking—we can still only infer why they typed what they
did. Theory cannot be abandoned. Moreover, the specific results cast into
doubt the free riding model where it’s wrongful for a non-trademark owner to
change consumer behavior regardless of what the consumer thinks—it seems that
unauthorized uses may not in fact work harm in the aggregate. However, in a big data world where we don’t
have any interest in knowing why behavior changed, only that it did, a court
might attempt to fine-tune the rule, and not allow keyword purchases on the naked mark or so called navigational
search, in order to make things even better for trademark owners. This is an example of a new legal theory that
can emerge from new methods of measurement; it wasn’t suggested by previous
doctrine. Whether that is a good or a
bad idea depends not on consumer behavior but on one’s theory of trademark
rights.
Lisa Oullette’s
paper on using search engines to make trademark judgments primarily addresses a
more basic question: how do we know whether a claimant has a mark at all, or a
strong mark. She argues that if a mark
is strong—either inherently distinctive or commercially strong—then many top
search results for that mark will relate to the source it identifies, so you
can use search engines to determine distinctiveness. Relatedly, she argues, the
extent of results overlap between searches for two different marks can also be
relevant for assessing the likelihood of confusion of those marks. For marks that don’t point uniquely to the
claimant except within a particular set of goods and services she suggests
using searches that add more information, like Mission Burritos rather than
Mission, which might be deemed non-navigational in the Bechtold/Tucker
framework.
Some thoughts:
(1) This theory may
have the most potential implications for registration system: registration
occurs in the abstract, not a full marketplace inquiry, and at least for word
marks does not concern itself with the current visual appearance of the
mark. The paper’s argument would imply
that we should take search engine results much more seriously in the
registration context to determine what a word means.
(2) However, we
still need a theory of trademark
meaning: example from the paper’s treatment of MICRO-THIN for condoms: all the
results for “micro thin condoms” referred to the relevant plaintiff’s mark, but
as the paper points out, this could mean strength for condoms or it could just
mean that the plaintiff is currently the only user of a not particularly
distinctive term within that category—I think this is a more significant
weakness than the paper acknowledges, because by entering “micro thin condoms”
as the search term you have already taken the position of a consumer using the
term to locate something, while what
we should want to know is “would a consumer use this term to identify
source”? The paper’s conclusion: “when
consumers search for MICROTHIN condoms, they are not simply looking for condoms
that are ‘extremely thin’—they are generally looking for Kimono MicroThin
condoms.” But that “when” contains the
assumption that drives the result. We
don’t know from these search results if the term is distinctive in the sense of
serving as an identifier of source—instead the paper’s inquiry is whether the
term is relatively unique within its field, and that really changes the basis
for trademark protection into a less consumer-focused rationale and far more
producer-oriented. The traditional
question of source significance asks “if” consumers search for micro-thin
condoms as an identifier of source, not what happens “when” they enter the
search terms. Questions I would be much
more interested in would be “how often do consumers search for Kimono condoms
or Kimono micro-thin condoms?” and “how often do consumers use micro-thin to
modify other brand names in their searches?”
The Bechtold/Tucker type of data are far more probative of that than the
search outputs.
(3) IP law’s largely
textual focus, which makes it much more confident handling words than other
symbols. Already reflected in the
doctrine: Abercrombie spectrum for
word marks might be empirically flawed, but it seems much more manageable than
alternatives proposed to identify inherent distinctiveness in symbols or trade
dress: Seabrook test for trade dress,
for example, is just another way of saying, four times, is this
distinctive? Unfortunately, just as with
current doctrine, image search is far behind word marks; the Google shortcut
might be most productive with respect to the least troublesome marks. And when there are both verbal and visual
components, there are additional complications: there are word marks that are
stronger when coupled with design features: a video “tube” site imitating
YouTube’s red and white ovals.
(4) Concerns about
using searches to assess likely confusion.
To take another example from the paper, TELMEX currently produces first
page results all related to one company, and the paper suggests that therefore
a new mark AUDITORIO TELMEX might be confusing.
But that seems to overweight the senior user’s interest: consider the
old Gruner
+ Jahr v. Meredith case, PARENTS versus PARENT’S DIGEST. When I type in “parents magazine” I get only
results for the former on the first page, but when I type in “parent’s digest”
I get no results for PARENTS magazine—one might conclude that PARENTS is a
rather weak mark since it doesn’t survive even minimal alteration.
The case law
suggests that we should be interested in what consumers think when they see the
junior use, and thus search on the junior
mark is more probative than search on the senior mark. When I type in
“Auditorio Telmex” I get no results relating to Telmex, the senior user. The results from the previous paper might
also be useful here, given that consumers often refine their searches when they
started with too abbreviated a term to get useful results. So, for example, if we found a certain number
of people trying Telmex first and then Auditorio Telmex, that could be evidence
not of confusion but of accurate understanding that these were two different
entities and that the initial shorthand just wasn’t enough to identify the
actual target. Note that we could call
that dilution, or we could say that such results show that Auditorio Telmex
wouldn’t dilute the naked Telmex mark at all, but we would need to be clear
about our definition of dilution to make that call.
[did not get to say
this para.] Relatedly: as price discrimination grows, different searchers may
get very different results depending on their previous purchases or demographic
profiles—which relates more closely to many of trademark’s concerns than
current geographic personalization: Northeastern
study showed this is already happening with segregation based on devices on
travel sites. Might support the paper’s
theory insofar as Google starts taking better account of differences in
consumers, so price points and consumer sophistication would actually divide
search results. On the other hand, the
paper’s argument about using Google to divine relatedness of goods or
likelihood of expansion does not seem persuasive: the paper says “if a
non-expanded mark is sufficiently strong that consumers might anticipate such
expansion, searches with keywords for those fields would likely still have
pages related to the mark,” but given search engines’ attempts to provide presently relevant results I can’t see
why that would be true—at the very least we’d need a lot more information about
who writes these pages saying “I can’t wait for the McDonald’s nav rattan
korma.”
These details
reinforce my concern that we should be very careful about when we change
doctrine in response to what we think we can measure. We might not even notice that we’re changing
the doctrine (which might, for example, narrow the protection for visual marks
if Google became a preferred source of evidence—I might be happy with that
outcome even if dubious about the mechanism).
We have a historical example of this evolution with judicial treatment
of survey evidence and the introduction of a requirement of control groups,
which had the practical effect of changing the percentage of consumers exposed
to the allegedly infringing use that courts recognized as sufficient to show
likely confusion. That’s probably the
happiest story of doctrinal evolution, but the decision to shape the doctrine
may be more normative than empirical, and that has to be kept in mind in all
these discussions.
Beebe: two cultures—marketers on one side and lawyers on the
other. Talking to marketers isn’t the same as talking to economists—who we talk
to affects the law just as talking to economists affects patent law. Marketers say: confusion as to source isn’t
interesting—let’s figure out how to measure dilution. Goes to slipperiness of models and constant
push of law away from confusion as to source and towards confusion as to
something else. Marketers: it’s not
consumers search costs, but TM search costs: TM law is currently dedicated to
minimizing the costs for TMs looking for consumers. Subjects are TMs/brands and
objects are consumers; lawyers don’t yet admit that.
Sprigman: for Steckel: cognitive delay as dilution? When I
get older, I get cognitively slower, have I diluted the mark? What’s the harm?
Steckel: when people get older: all associations are
delayed, that’s what a control group is for.
There is research showing that even for the most important of
purchases—home, car, etc., that at any moment only 3-4 attributes are
operative. People have limited info processing capabilities and as such you
want to put your best pitch up front.
New Coke debacle: New Coke tasted better, but people didn’t buy the
product for its taste.
Sprigman: but there’s no falsity injected into associations
through “dilution.” What you’ve posited,
is substantive weakening of meaning, can see effect of that, but not new
associations. Tarnishment story makes
more sense, but is a huge First Amendment problem since in every other context
you’re allowed to say “Victoria’s Secret sucks,” even as a competitor.
Steckel: you don’t buy that measuring reaction times
captures notion of impairing distinctiveness.
Sprigman: need relevance in retail environment.
Steckel: by delaying reaction times, it bumps something from
short term memory to long term memory, taking longer to evoke in world of
limited attention.
McKenna: this is two cultures: lawyer wants to know where
the associations are, with source 1 or source 2. Steckel is asking what associations will
arise—that’s a mismatch. Substantive
meaning of mark v. whether it will take longer to know whether I’m dealing with
New Coke or Old Coke.
Sprigman: whose views count for the law? Typically we don’t impinge freedom of speech
without harm.
RT: I wrote a
whole article about this. Don’t
agree with the characterization that delays correspond to shift from short term
memory to long term memory. 160
milliseconds is not a long enough delay.
This is an example of what happens to empirical work laundered by
lawyers: it gets ramped up as more significant than it is in the translation;
we regularly mistake statistical significance for practical significance.
Bechtold: Use of big data and limitations. One response: need more data. In some areas, this is true—initial interest
confusion; we tried to test this and could’ve tested this w/right data. In general, we are only slowly seeing usable
data. Fully agree that this is not a
substitute for a theory of TM that is justified. Studies can help us find new theories/new
effects. Spillover effects from
unauthorized uses can benefit brand owner; no one has clear theoretical view of
how this works psychologically or what TM lawyers should do w/ it.
Jeremy Sheff: should ask what questions empirical methods
should be used to answer and that depends on our theory of TM law. Consumer
psychology: what are we trying to measure in the minds of consumers? Marketing departments have been looking at
these questions though not attuned to doctrine. Competition: if we think there
are empirical questions, we could turn to competition/antitrust law. Look at effects on entry, prices, output,
elasticity of demand. These aren’t being investigated right now. That could
help make a case for a particular theory or explore our theoretical
presumptions.
Orly Lobel: useful to think about this panel v. patents
panel. Much has to do with whether we’re
having the same conversations: does the theory map onto the work? Confusion
seems like an instrumental, narrow goal—could it be consumer welfare, new
entry, etc.? Current work seems
defeatist, focused on what we have now on the books. This panel spoke to adjudicators: how to apply/interpret
doctrinal mandate; previous panel was willing to talk about what the law should
be. Maybe we just don’t know how to
aggregate the effects of TMs. Or maybe
we’ve seen more action on patent reform and thus had more demand for studies of
bigger Qs.
McKenna: patent traditionally wanted to promote more output.
TM traditionally just wanted to prohibit particular practices. Yet TM is used as a substitute for or
complement to patent law; can be measured in terms of same effects. Antitrust people: don’t think TM has any
effect on competition/any market power for brands. That’s bound up in history,
but we need more study.
Chris Buccafusco: TM doesn’t start w/empirical assumptions
about brains and decisions and then try to figure out harms; it starts
w/assumptions about what bads are and tries to get evidence from social
scientists. Mental map of associations
is starting point of marketers, but not TM.
Might be any number of conceptions of how consumers use brands; could be
less cognitivist models. Might produce
different sets of harms. His skepticism:
TM owners want to find all the harms and use all possible theories of
decisionmaking as long as they specify potentially compensable harms.
Gregory Mandel: move to the ought is useful; so is how the
law is working on the ground, as the TM papers were focused on.
Scott Hemphill: clickstream data: you can see sometimes that
people stayed only for 5-7 seconds. That’s a signal they didn’t get what they
wanted. Could also get that from sequence of queries. Might be a way to detect
brief confusion (though, RT notes, also self-refutes theory of harm to TM
owners).
Irina Manta: not sure that association w/porn is the only
way or even the most direct way to get at tarnishment. If you believe in the
model, Victor’s Little Secret creates negative association, but people don’t
remember why they have a negative association when they see VS again, it’s just
a generalized dis-ease. Quality
perception reduced; that should be actionable every time there’s no First
Amendment protection, according to TM owners.
(Of course, under Central Hudson,
“every time there’s no First Amendment protection” means “every time there is a
factually false statement about the TM owner in commercial speech,” which
tarnishment is not.)
Steckel: hard to expose respondents to an experience that
mimicks what happens in the market; simply asking them about associations
w/porn may not do it.
Michael Meurer: there have been some market value studies
about value of TMs. Not sure whether there are studies on market effects of TM
litigation, but there could be.
McKenna: there are European economists who try to study
relations to innovation, but there are many problems. First, they just measure
TM applications which has correlation/causation problems.
Katherine Strandburg: Papers related to Abercrombie emphasized the importance of context in what consumers
see. Google genericity case: court
distinguished verb use from TM use.
People use google generically as a verb but not as a source. Can we split the baby here, giving people
more protection for the mark in a certain context, but less in other contexts
where the use isn’t the same? Also,
tension in Oullette paper: if you do a search and the first page is all about a
particular TM, then she argues it’s distinctiveness. But wouldn’t that also
suggest a lack of confusion? That seems
perfectly sensible—a more distinctive mark might be less confusable.
Barton Beebe: TM says that stronger mark = greater scope of
protection/more likely confusion is. If we abandoned that principle, the
edifice would crumble. Some European
courts have suggested that strength decreases likely confusion, but not
American courts. (RT: I would say that individual US courts have occasionally
reinvented this theory in parody/First Amendment cases, but haven’t made it
coherent.)
McKenna: not about empirical reality, b/c empirical evidence
supports Strandburg.
Brett Frischmann: Interesting empirical questions would
challenge assumptions/theory of TM.
Assumptions about stable/fixed preferences of consumers; examine way TMs
enable producers to shape consumer preferences. What are the public harms from
TM? Deadweight loss—how to measure in TM law? Chilling effects, speech values.
No comments:
Post a Comment