instruction
stringlengths 93
525
| input
stringlengths 121
36.8k
| output
stringlengths 100
8.28k
| __index_level_0__
int64 0
1.92k
| messages
listlengths 3
3
|
---|---|---|---|---|
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: 2. It adequately compensates end-users.
Highlight Ratio: 0.2 | A new paradigm of civil compensation for damages related to AI: towards the evolution of compensation from an issue of civil liability to one of financial management of losses
The law,
binds economic and social activities in order to contribute to the pursuit of welfare; on the other hand, however, the law cannot arbitrarily define its objectives and (especially) the means. The actual functioning of the economic and social contexts faced must be taken into the utmost consideration
in order to develop well-founded, affordable, reliable and effective rules (de Jong et al. 2018).
The failure of the current paradigm of civil liability based on deterrence, when applied to artificial intelligence, observed and (I believe) established above, requires a radical modification thereof. Such a modification appears relevant in these days, since the application of the “traditional” paradigm of civil liability can hinder the development of markets towards the intensive use of artificial intelligence and robotisation in the future (the
technology chilling
Furthermore, civil liability rules rooted in deterrence are likely to place jurisdictions adopting this paradigm at a competitive disadvantage in favour of jurisdictions that are more responsive to the needs and demands of the markets referred to.
We need to adapt the legislation to the “risk society”, that is: “a systematic way of dealing with hazards and insecurities induced and introduced by modernization
Such a conclusion should lead to discarding the “blame culture”, which inspires and supports the current law on civil liability, and replacing it, at least in some cases (
b
with a “no-blame culture”, rooted in risk management
the negative externalities imposed on the AI markets by the traditional civil liability paradigm could be reduced if producers and programmers of artificial intelligence devices could be released from civil liability under certain conditions; in particular, when there is no evidence of their negligence, imprudence or unskillfulness and their activity complied with scientifically validated standards
Such release, however, may not (and should not) lead to prevent damaged customers and end-users to get compensation
This is why I believe that a new regulation of the matter should be developed, inspired by a new paradigm, aimed at maintaining compensation for damages on the patient’s side, but shifting away from producers and programmers of AI devices
This new paradigm could be built on the basis of “no-fault” systems available in different jurisdictions
one can cite the no-fault rules issued in the field of medical damage
adverse effects attributed to vaccination
damages coming from unknown drivers
Adopting a “no-fault” scheme would isolate compensation in favour of damaged end-users from liability on producers and programmers of AI devices. It would also help resolve other weaknesses inherent in the traditional paradigm of civil liability
One can mention, here, the risk of civil liability turning into a “damages lottery” due to the fact that, in some cases, the damages cannot be awarded because no one is at fault in the specific event. It is also possible to report the case in which damages cannot be collected because the debtor is
unable to pay | We need to adapt
to
risk society
to discard
blame culture
which inspires
civil liability, and replac
it
with
no-blame
risk management
negative externalities imposed on
AI markets by
civil liability
could be reduced if producers
could be released from
liability
when there is no ev
of
negligence
Such release
should not
prevent damaged customers
to get compensation
a new regulation
should be developed,
on the basis of “no-fault” systems
one can cite
no-fault
in the field of medical damage
adverse effects attributed to vaccination
damages coming from unknown drivers
no-fault
would isolate compensation in favour of damaged end-users from liability on producers | 211 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: 2. It adequately compensates end-users.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "A new paradigm of civil compensation for damages related to AI: towards the evolution of compensation from an issue of civil liability to one of financial management of losses\nThe law,\nbinds economic and social activities in order to contribute to the pursuit of welfare; on the other hand, however, the law cannot arbitrarily define its objectives and (especially) the means. The actual functioning of the economic and social contexts faced must be taken into the utmost consideration\nin order to develop well-founded, affordable, reliable and effective rules (de Jong et al. 2018).\nThe failure of the current paradigm of civil liability based on deterrence, when applied to artificial intelligence, observed and (I believe) established above, requires a radical modification thereof. Such a modification appears relevant in these days, since the application of the “traditional” paradigm of civil liability can hinder the development of markets towards the intensive use of artificial intelligence and robotisation in the future (the\ntechnology chilling\nFurthermore, civil liability rules rooted in deterrence are likely to place jurisdictions adopting this paradigm at a competitive disadvantage in favour of jurisdictions that are more responsive to the needs and demands of the markets referred to.\nWe need to adapt the legislation to the “risk society”, that is: “a systematic way of dealing with hazards and insecurities induced and introduced by modernization\nSuch a conclusion should lead to discarding the “blame culture”, which inspires and supports the current law on civil liability, and replacing it, at least in some cases (\nb\nwith a “no-blame culture”, rooted in risk management\nthe negative externalities imposed on the AI markets by the traditional civil liability paradigm could be reduced if producers and programmers of artificial intelligence devices could be released from civil liability under certain conditions; in particular, when there is no evidence of their negligence, imprudence or unskillfulness and their activity complied with scientifically validated standards\nSuch release, however, may not (and should not) lead to prevent damaged customers and end-users to get compensation\nThis is why I believe that a new regulation of the matter should be developed, inspired by a new paradigm, aimed at maintaining compensation for damages on the patient’s side, but shifting away from producers and programmers of AI devices\nThis new paradigm could be built on the basis of “no-fault” systems available in different jurisdictions\none can cite the no-fault rules issued in the field of medical damage\nadverse effects attributed to vaccination\ndamages coming from unknown drivers\nAdopting a “no-fault” scheme would isolate compensation in favour of damaged end-users from liability on producers and programmers of AI devices. It would also help resolve other weaknesses inherent in the traditional paradigm of civil liability\nOne can mention, here, the risk of civil liability turning into a “damages lottery” due to the fact that, in some cases, the damages cannot be awarded because no one is at fault in the specific event. It is also possible to report the case in which damages cannot be collected because the debtor is\nunable to pay",
"role": "user"
},
{
"content": "We need to adapt\nto\nrisk society\nto discard\nblame culture\nwhich inspires\ncivil liability, and replac\nit\nwith\nno-blame\nrisk management\nnegative externalities imposed on\nAI markets by\ncivil liability\ncould be reduced if producers\ncould be released from\nliability\nwhen there is no ev\nof\nnegligence\nSuch release\nshould not\nprevent damaged customers\nto get compensation\na new regulation\nshould be developed,\non the basis of “no-fault” systems\none can cite\nno-fault\nin the field of medical damage\nadverse effects attributed to vaccination\ndamages coming from unknown drivers\nno-fault\nwould isolate compensation in favour of damaged end-users from liability on producers",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t. | The divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person, once it has decided to recognize the legal person as such. We should resolve the issue of the legal personhood of robots at this level, rather than treating legal personhood as an all-or-nothing black box (Koops et al. 2010, p. 556). Edsger Dijkstra has noted, “A convincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box” (Dijkstra 1970). A legal system, if it chose to confer legal personality on robots, would need to say specifically which legal rights and obligations went with the designation. If it does not, then the legal system will struggle, as happened with the Bank for International Settlements, to make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box. In line with the fictionalist paradigm, and as the ICJ opined with respect to the UN, the legal system should determine the legal rights and obligations of a new legal person by reference to how the legal person relates to the legal system’s purposes. | divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person
We should resolve the issue of the legal personhood
at this level, rather than treating legal personhood as an all-or-nothing black box
convincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box
A legal system, if it chose to confer legal personality
would need to say specifically which legal rights and obligations went with the designation. If it does not
the legal system will struggle
to make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box | 212 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t.",
"role": "system"
},
{
"content": "The divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person, once it has decided to recognize the legal person as such. We should resolve the issue of the legal personhood of robots at this level, rather than treating legal personhood as an all-or-nothing black box (Koops et al. 2010, p. 556). Edsger Dijkstra has noted, “A convincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box” (Dijkstra 1970). A legal system, if it chose to confer legal personality on robots, would need to say specifically which legal rights and obligations went with the designation. If it does not, then the legal system will struggle, as happened with the Bank for International Settlements, to make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box. In line with the fictionalist paradigm, and as the ICJ opined with respect to the UN, the legal system should determine the legal rights and obligations of a new legal person by reference to how the legal person relates to the legal system’s purposes.",
"role": "user"
},
{
"content": "divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person\nWe should resolve the issue of the legal personhood\nat this level, rather than treating legal personhood as an all-or-nothing black box\nconvincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box\nA legal system, if it chose to confer legal personality\nwould need to say specifically which legal rights and obligations went with the designation. If it does not\nthe legal system will struggle\nto make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t. | divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person
We should resolve the issue of the legal personhood
at this level, rather than treating legal personhood as an all-or-nothing black box
convincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box
A legal system, if it chose to confer legal personality
would need to say specifically which legal rights and obligations went with the designation. If it does not
the legal system will struggle
to make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box | divisibility
which rights and duties
at this level
impossible
only hope
would need to say specifically which legal rights and obligations
struggle
to make sense of what it has done
without being more specific | 213 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t.",
"role": "system"
},
{
"content": "divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person\nWe should resolve the issue of the legal personhood\nat this level, rather than treating legal personhood as an all-or-nothing black box\nconvincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box\nA legal system, if it chose to confer legal personality\nwould need to say specifically which legal rights and obligations went with the designation. If it does not\nthe legal system will struggle\nto make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box",
"role": "user"
},
{
"content": "divisibility\nwhich rights and duties\nat this level\nimpossible\nonly hope\nwould need to say specifically which legal rights and obligations\nstruggle\nto make sense of what it has done\nwithout being more specific",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t.
Highlight Ratio: 0.5 | divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person
We should resolve the issue of the legal personhood
at this level, rather than treating legal personhood as an all-or-nothing black box
convincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box
A legal system, if it chose to confer legal personality
would need to say specifically which legal rights and obligations went with the designation. If it does not
the legal system will struggle
to make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box | divisibility
raises
question
which rights and duties
system should confer
We should resolve
legal personhood
at this level, rather than treating
personhood as
all-or-nothing
legal system
would need to say specifically which
rights and obligations
If
not
the
system will struggle
to make sense of what it has done. To
confer “legal personality,” without being more specific, is to
legal personality as a black box | 214 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Vesting rights and duties in new legal personalities requires specifying the exact rights and obligations---the CP doesn’t.\n\nHighlight Ratio: 0.5",
"role": "system"
},
{
"content": "divisibility of legal personhood raises the question of which rights and duties a legal system should confer on a legal person\nWe should resolve the issue of the legal personhood\nat this level, rather than treating legal personhood as an all-or-nothing black box\nconvincing demonstration of correctness being impossible as long as the mechanism is regarded as a black box, our only hope lies in not regarding the mechanism as a black box\nA legal system, if it chose to confer legal personality\nwould need to say specifically which legal rights and obligations went with the designation. If it does not\nthe legal system will struggle\nto make sense of what it has done. To try to confer “legal personality,” without being more specific, is to regard legal personality as a black box",
"role": "user"
},
{
"content": "divisibility\nraises\nquestion\nwhich rights and duties\nsystem should confer\nWe should resolve\nlegal personhood\nat this level, rather than treating\npersonhood as\nall-or-nothing\nlegal system\nwould need to say specifically which\nrights and obligations\nIf\nnot\nthe\nsystem will struggle\nto make sense of what it has done. To\nconfer “legal personality,” without being more specific, is to\nlegal personality as a black box",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Vesting language has to be explicit. | a. Statements that Benefits "May" Vest
Ostensibly to argue for implied vesting, Plaintiffs emphasize that at several different points, ASARCO or ASARCO negotiators represented that retiree healthcare benefits "may" be vested. (E.g. Resp. at 19; Resp. SOF at 22-23, 40, 47-48.) But this evidence fails to establish that a reasonable factfinder could determine that lifetime benefits were vested for the Class Members. At most, Plaintiffs have established that certain extrinsic evidence suggests it was unclear whether the benefits were vested. However, echoing [*33] ERISA's promotion of employer flexibility regarding provision of welfare benefits, Tackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits. See Tackett, 574 U.S. at 442; Reese, 138 S. Ct. at 763-64 (condemning inference that benefits vest when contract is silent as to duration of benefits). On the contrary, Tackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits, and permissive language reflects no clear intent to vest. See Tackett, 574 U.S. at 435, 441; Reese, 138 S. Ct. at 765. | Statements that Benefits "May" Vest
Ostensibly to argue for implied vesting, Plaintiffs emphasize
ASARCO
represented
retiree healthcare benefits "may" be vested
this evidence fails to establish that
benefits were vested
Plaintiffs have established that
evidence suggests it was unclear whether the benefits were vested
Tackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.
Tackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits | 215 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Vesting language has to be explicit.",
"role": "system"
},
{
"content": "a. Statements that Benefits \"May\" Vest\nOstensibly to argue for implied vesting, Plaintiffs emphasize that at several different points, ASARCO or ASARCO negotiators represented that retiree healthcare benefits \"may\" be vested. (E.g. Resp. at 19; Resp. SOF at 22-23, 40, 47-48.) But this evidence fails to establish that a reasonable factfinder could determine that lifetime benefits were vested for the Class Members. At most, Plaintiffs have established that certain extrinsic evidence suggests it was unclear whether the benefits were vested. However, echoing [*33] ERISA's promotion of employer flexibility regarding provision of welfare benefits, Tackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits. See Tackett, 574 U.S. at 442; Reese, 138 S. Ct. at 763-64 (condemning inference that benefits vest when contract is silent as to duration of benefits). On the contrary, Tackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits, and permissive language reflects no clear intent to vest. See Tackett, 574 U.S. at 435, 441; Reese, 138 S. Ct. at 765.",
"role": "user"
},
{
"content": "Statements that Benefits \"May\" Vest\nOstensibly to argue for implied vesting, Plaintiffs emphasize\nASARCO\nrepresented\nretiree healthcare benefits \"may\" be vested\nthis evidence fails to establish that\nbenefits were vested\nPlaintiffs have established that\nevidence suggests it was unclear whether the benefits were vested\nTackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.\nTackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Vesting language has to be explicit. | Statements that Benefits "May" Vest
Ostensibly to argue for implied vesting, Plaintiffs emphasize
ASARCO
represented
retiree healthcare benefits "may" be vested
this evidence fails to establish that
benefits were vested
Plaintiffs have established that
evidence suggests it was unclear whether the benefits were vested
Tackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.
Tackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits | "May" Vest
implied vesting
"may" be vested
were vested
unclear
benefits were vested
Tackett
Reese
presumption
benefits vest
employer's intention
vest benefits
mandate
vesting
clear intent
vest benefits | 216 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Vesting language has to be explicit.",
"role": "system"
},
{
"content": "Statements that Benefits \"May\" Vest\nOstensibly to argue for implied vesting, Plaintiffs emphasize\nASARCO\nrepresented\nretiree healthcare benefits \"may\" be vested\nthis evidence fails to establish that\nbenefits were vested\nPlaintiffs have established that\nevidence suggests it was unclear whether the benefits were vested\nTackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.\nTackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits",
"role": "user"
},
{
"content": "\"May\" Vest\nimplied vesting\n\"may\" be vested\nwere vested\nunclear\nbenefits were vested\nTackett\nReese\npresumption\nbenefits vest\nemployer's intention\nvest benefits\nmandate\nvesting\nclear intent\nvest benefits",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Vesting language has to be explicit.
Highlight Ratio: 0.4 | Statements that Benefits "May" Vest
Ostensibly to argue for implied vesting, Plaintiffs emphasize
ASARCO
represented
retiree healthcare benefits "may" be vested
this evidence fails to establish that
benefits were vested
Plaintiffs have established that
evidence suggests it was unclear whether the benefits were vested
Tackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.
Tackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits | implied vesting
fails to establish
benefits were vested
Tackett
forbid
applying a presumption that benefits vest absent
intention to vest
Tackett
mandate
vesting does not occur without
clear intent to vest benefits | 217 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Vesting language has to be explicit.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "Statements that Benefits \"May\" Vest\nOstensibly to argue for implied vesting, Plaintiffs emphasize\nASARCO\nrepresented\nretiree healthcare benefits \"may\" be vested\nthis evidence fails to establish that\nbenefits were vested\nPlaintiffs have established that\nevidence suggests it was unclear whether the benefits were vested\nTackett and Reese forbid any adjudicator from applying a presumption that benefits vest absent evidence of an employer's intention to vest benefits.\nTackett and Reese mandate that vesting does not occur without the parties' clear intent to vest benefits",
"role": "user"
},
{
"content": "implied vesting\nfails to establish\nbenefits were vested\nTackett\nforbid\napplying a presumption that benefits vest absent\nintention to vest\nTackett\nmandate\nvesting does not occur without\nclear intent to vest benefits",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process. | When in the debate about the possible humanness of computers the MIT computer scientist Joseph Weizenbaum was hard pressed to concede that computers can be “socialized”, he found that they could but in an extremely limited sense. He then stated:
If both machines and humans are socializable, then we must ask in what way the socialization of the human must necessarily be different from that of the machine. The answer is, of course, so obvious that it makes the very asking of the question appear ludicrous, if indeed not obscene. It is a sign of the madness of our time that this issue is to be addressed at all. (Weizenbaum, 1976, p. 210)
For Weizenbaum the answer was so obvious that he even did not bother to give it. He probably thought that, as science progressed, the answer would eventually become obvious to everyone in the field anyway. But this did not happen. The opposite happened. Today, nearly half a century later, it is common “knowledge” among scientists and in a wide public that computers will, sooner or later, be human-like. Some think that computers already are.
So was Weizenbaum eventually proven wrong? Was he, who reasoned about the future of the computer world, about the “digital age”, as thoroughly as nobody else, unable to foresee a future where the difference between humans and machines would become blurred, as many in the AI community want to make us believe, a development which, if it actually happened, would have a transforming impact on the future of mankind? In the decades following his above statement, Weizenbaum actually never distanced himself from it. Quite to the contrary, he was more and more convinced that a serious misdevelopment was occuring in man’s attitude toward computers. So before we discuss his decades old statement as one caused by lack of fantasy and made obsolete by scientific progress, we had better take pains to delve into the scientific-philosophical depths of the man-machine issue. This is what this book is about. Unfortunately, this issue is not only, and predictably so, an extremely complex one, but, more importantly, the whole debate is also plagued by a deep ignorance about the nature of human mental processes, about the human psyche. This need not surprise us, because the debate is largely dominated by engineers, mathematicians and information scientists. Sadly, psychologists, actually the specialists responsible for the very mental functions the computer scientists intend to emulate, have contributed little to instructing those computer scientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human-like wave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or even superhuman, and that they eventually will take over control of the world from those outmoded creatures that used to be called humans.
[INTRODUCTION BEGINS]
When planning this book, in which I want to refute the idea that there can be processes in machines comparable to mental processes in humans, i.e. that machines can be human-like, I first thought to take typical examples of computer functions allegedly comparable to human mental functions, compare the two and then decide whether it makes sense to regard computer functions as really human-like. But I soon realized that my task would be comparable to Sisyphus’. With the latest computer accomplishment popping up on a monthly if not weekly basis, I would have to start pushing the rock uphill anew every month or week with no end in sight. The book would have to be rewritten at least once a year. Machines with human-like mental functions now have been predicted for many decades (since the advent of electronic computers) and I am afraid that this will go on for many decades more unless the scientific community realizes that there are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons (as with, for example, intelligence) to make no sense whatsoever. So the book will be about a priori grounds on which computers cannot be human-like.
When, long ago, I read Weizenbaum’s comment that to compare the socialization of a machine with that of a human is a sign of madness, I immediately felt that this was the ultimate comment possible and did not expect Weizenbaum to explain why (which he did not bother to do anyway). But the insight into this madness not having befallen the scientific community nearly half a century later that “Why” must finally be delivered.
A comparison between the socialization of a human and that of a computer is not seen as mad in the AI community. This is so not because obvious similarities could be shown to exist between the two (there aren’t any; there is just a faint analogy in that both are somehow affected by their environment) but simply because this kind of speak (and think) has become quite common among AI researchers. Take a typical phrase like “culture is poured into artificial brains” (Collins, 2018, p. 173). You can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely, it may not be felt as nonsense any more. The above phrase about how culture can manage to get into a machine reminds one in its nonsensicality of Francis Picabia’s1 “The head is round in order to allow the thoughts to change direction”. At a closer look much of AI speak appears to be inspired by Dadaism. The only difference between AI and Dada speak is that Dada is funny and the Dadaists consciously invented nonsensicalities in order to provoke and entertain, while AI speak is by no means funny and the AI researchers are convinced that they talk reasonably. So when reading the AI literature, we are confronted with all kinds of absurdities, like machines being socialized, having human feelings, motives, consciousness and religion, westernized human-like desires, and culture being poured or fed into machines, or with artificial slaves with human-like bodies, or with self replicating intelligent machines. I think that large parts of the AI literature may be seen as lessons in absurdity, presented as rational, scientifically based predictions about our digital future.
The idea of human-like computers is based, quite like its 18th century predecessor, L’homme machine (Man a Machine) by de la Mettrie (1921), on a simple mechanistic-materialistic belief, abandoned in physics long ago. As a consequence of this simplistic man-a-machine view, the whole debate about human mental functions in computers suffers, as mentioned, from a deep ignorance about the nature of human mental processes, about the human psyche. In addition to that, the engineering perspective of the field largely underestimates our deep ignorance about the biological, largely neural, basis (the biological substrate) of those processes. So I will in some detail deal with that substrate and then try to paint a realistic picture of the daunting complexity of human mental functions in accord with the simple motto “if you want to mimic something you should know what it is.”
That picture will constitute the largest part of the book. My view of the man-a-machine matter is not just that the idea is a somehow problematic one, but that it is an absurd, i.e. a ridiculousy nonsensical one. This view, identical to the one Weizenbaum took half a century ago, today is one held by a small minority facing a huge majority of believers in human-like computers, both in a wide public and also among scientists. To convince the reader to join a small minority against a seemingly overwhelming opposition is no easy task. As a kind of psychological support in this task let me, before coming to the actual topic of the book (the absurdity of the man-a-machine notion), present some examples from mental history where a large majority had it wrong for decades, centuries or even milennia. And let me also point to some peculiar characteristics of the debate which must make us doubt that we are dealing with a scientific one in the first place. In terms of argumentation, the field must be seen as one following entirely its own rules. | we must ask in what way
the human must necessarily be different from
the machine. The answer is
so obvious
it makes the very asking of the question appear ludicrous, if
not obscene. It is a sign of
madness
that this
is to be addressed at all
Weizenbaum
probably thought
as science progressed, the answer would
become obvious to everyone in the field
this did not happen. The opposite happened
Today
it is common “knowledge”
that computers will, sooner or later, be human-like
So was Weizenbaum
proven wrong?
Quite
the contrary
the debate is largely dominated by engineers, mathematicians and info
scientists
psychologists
the specialists responsible for the
mental functions the computer scientists intend to emulate
contributed little to instructing those
scientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human
wave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or
superhuman, and that they eventually will take over control of the world
I want to refute the idea that there can be
machines comparable to
humans
I soon realized that my task would be comparable to Sisyphus’
With the latest
accomplishment popping up
weekly
I would have to
push
the rock
every
week with no end in sight
The book would have to be rewritten
once a year. Machines with human-like
functions
have been predicted for
decades
this will go on for
decades more unless the scientific community realizes
there are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons
to make no sense whatsoever
A comparison between
human and
computer is not seen as mad in the AI community
not because obvious similarities could be shown to exist
there aren’t any
but
because this kind of speak
has become
common among AI researchers. Take a typical phrase like “culture is poured into artificial brains”
You can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely
may not be felt as nonsense any more. The
phrase about how culture can
get into a machine reminds one in its nonsensicality of
The head is round
to allow
thoughts to change direction
much of AI speak appears
inspired by Dadaism. The only difference between AI and Dada
is that Dada is funny and the Dadaists consciously invented nonsensicalities
while AI
is by no means funny and
AI researchers are convinced
they talk reasonably
when reading the AI literature, we are confronted with all kinds of absurdities, like machines
having
motives
westernized
desires
or
human-like bodies, or
self replicating intelligent machines
the engineering perspective of the field largely underestimates our deep ignorance about the biological
basis
of those processes
“if you want to mimic something you should know what it is
the man-a-machine
is not just
somehow problematic
but
it is
absurd
ridiculousy nonsensical | 218 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process.",
"role": "system"
},
{
"content": "When in the debate about the possible humanness of computers the MIT computer scientist Joseph Weizenbaum was hard pressed to concede that computers can be “socialized”, he found that they could but in an extremely limited sense. He then stated:\nIf both machines and humans are socializable, then we must ask in what way the socialization of the human must necessarily be different from that of the machine. The answer is, of course, so obvious that it makes the very asking of the question appear ludicrous, if indeed not obscene. It is a sign of the madness of our time that this issue is to be addressed at all. (Weizenbaum, 1976, p. 210)\nFor Weizenbaum the answer was so obvious that he even did not bother to give it. He probably thought that, as science progressed, the answer would eventually become obvious to everyone in the field anyway. But this did not happen. The opposite happened. Today, nearly half a century later, it is common “knowledge” among scientists and in a wide public that computers will, sooner or later, be human-like. Some think that computers already are.\nSo was Weizenbaum eventually proven wrong? Was he, who reasoned about the future of the computer world, about the “digital age”, as thoroughly as nobody else, unable to foresee a future where the difference between humans and machines would become blurred, as many in the AI community want to make us believe, a development which, if it actually happened, would have a transforming impact on the future of mankind? In the decades following his above statement, Weizenbaum actually never distanced himself from it. Quite to the contrary, he was more and more convinced that a serious misdevelopment was occuring in man’s attitude toward computers. So before we discuss his decades old statement as one caused by lack of fantasy and made obsolete by scientific progress, we had better take pains to delve into the scientific-philosophical depths of the man-machine issue. This is what this book is about. Unfortunately, this issue is not only, and predictably so, an extremely complex one, but, more importantly, the whole debate is also plagued by a deep ignorance about the nature of human mental processes, about the human psyche. This need not surprise us, because the debate is largely dominated by engineers, mathematicians and information scientists. Sadly, psychologists, actually the specialists responsible for the very mental functions the computer scientists intend to emulate, have contributed little to instructing those computer scientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human-like wave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or even superhuman, and that they eventually will take over control of the world from those outmoded creatures that used to be called humans.\n[INTRODUCTION BEGINS]\nWhen planning this book, in which I want to refute the idea that there can be processes in machines comparable to mental processes in humans, i.e. that machines can be human-like, I first thought to take typical examples of computer functions allegedly comparable to human mental functions, compare the two and then decide whether it makes sense to regard computer functions as really human-like. But I soon realized that my task would be comparable to Sisyphus’. With the latest computer accomplishment popping up on a monthly if not weekly basis, I would have to start pushing the rock uphill anew every month or week with no end in sight. The book would have to be rewritten at least once a year. Machines with human-like mental functions now have been predicted for many decades (since the advent of electronic computers) and I am afraid that this will go on for many decades more unless the scientific community realizes that there are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons (as with, for example, intelligence) to make no sense whatsoever. So the book will be about a priori grounds on which computers cannot be human-like.\nWhen, long ago, I read Weizenbaum’s comment that to compare the socialization of a machine with that of a human is a sign of madness, I immediately felt that this was the ultimate comment possible and did not expect Weizenbaum to explain why (which he did not bother to do anyway). But the insight into this madness not having befallen the scientific community nearly half a century later that “Why” must finally be delivered.\nA comparison between the socialization of a human and that of a computer is not seen as mad in the AI community. This is so not because obvious similarities could be shown to exist between the two (there aren’t any; there is just a faint analogy in that both are somehow affected by their environment) but simply because this kind of speak (and think) has become quite common among AI researchers. Take a typical phrase like “culture is poured into artificial brains” (Collins, 2018, p. 173). You can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely, it may not be felt as nonsense any more. The above phrase about how culture can manage to get into a machine reminds one in its nonsensicality of Francis Picabia’s1 “The head is round in order to allow the thoughts to change direction”. At a closer look much of AI speak appears to be inspired by Dadaism. The only difference between AI and Dada speak is that Dada is funny and the Dadaists consciously invented nonsensicalities in order to provoke and entertain, while AI speak is by no means funny and the AI researchers are convinced that they talk reasonably. So when reading the AI literature, we are confronted with all kinds of absurdities, like machines being socialized, having human feelings, motives, consciousness and religion, westernized human-like desires, and culture being poured or fed into machines, or with artificial slaves with human-like bodies, or with self replicating intelligent machines. I think that large parts of the AI literature may be seen as lessons in absurdity, presented as rational, scientifically based predictions about our digital future.\nThe idea of human-like computers is based, quite like its 18th century predecessor, L’homme machine (Man a Machine) by de la Mettrie (1921), on a simple mechanistic-materialistic belief, abandoned in physics long ago. As a consequence of this simplistic man-a-machine view, the whole debate about human mental functions in computers suffers, as mentioned, from a deep ignorance about the nature of human mental processes, about the human psyche. In addition to that, the engineering perspective of the field largely underestimates our deep ignorance about the biological, largely neural, basis (the biological substrate) of those processes. So I will in some detail deal with that substrate and then try to paint a realistic picture of the daunting complexity of human mental functions in accord with the simple motto “if you want to mimic something you should know what it is.”\nThat picture will constitute the largest part of the book. My view of the man-a-machine matter is not just that the idea is a somehow problematic one, but that it is an absurd, i.e. a ridiculousy nonsensical one. This view, identical to the one Weizenbaum took half a century ago, today is one held by a small minority facing a huge majority of believers in human-like computers, both in a wide public and also among scientists. To convince the reader to join a small minority against a seemingly overwhelming opposition is no easy task. As a kind of psychological support in this task let me, before coming to the actual topic of the book (the absurdity of the man-a-machine notion), present some examples from mental history where a large majority had it wrong for decades, centuries or even milennia. And let me also point to some peculiar characteristics of the debate which must make us doubt that we are dealing with a scientific one in the first place. In terms of argumentation, the field must be seen as one following entirely its own rules.",
"role": "user"
},
{
"content": "we must ask in what way\nthe human must necessarily be different from\nthe machine. The answer is\nso obvious\nit makes the very asking of the question appear ludicrous, if\nnot obscene. It is a sign of\nmadness\nthat this\nis to be addressed at all\nWeizenbaum\nprobably thought\nas science progressed, the answer would\nbecome obvious to everyone in the field\nthis did not happen. The opposite happened\nToday\nit is common “knowledge”\nthat computers will, sooner or later, be human-like\nSo was Weizenbaum\nproven wrong?\nQuite\nthe contrary\nthe debate is largely dominated by engineers, mathematicians and info\nscientists\npsychologists\nthe specialists responsible for the\nmental functions the computer scientists intend to emulate\ncontributed little to instructing those\nscientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human\nwave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or\nsuperhuman, and that they eventually will take over control of the world\nI want to refute the idea that there can be\nmachines comparable to\nhumans\nI soon realized that my task would be comparable to Sisyphus’\nWith the latest\naccomplishment popping up\nweekly\nI would have to\npush\nthe rock\nevery\nweek with no end in sight\nThe book would have to be rewritten\nonce a year. Machines with human-like\nfunctions\nhave been predicted for\ndecades\nthis will go on for\ndecades more unless the scientific community realizes\nthere are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons\nto make no sense whatsoever\nA comparison between\nhuman and\ncomputer is not seen as mad in the AI community\nnot because obvious similarities could be shown to exist\nthere aren’t any\nbut\nbecause this kind of speak\nhas become\ncommon among AI researchers. Take a typical phrase like “culture is poured into artificial brains”\nYou can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely\nmay not be felt as nonsense any more. The\nphrase about how culture can\nget into a machine reminds one in its nonsensicality of\nThe head is round\nto allow\nthoughts to change direction\nmuch of AI speak appears\ninspired by Dadaism. The only difference between AI and Dada\nis that Dada is funny and the Dadaists consciously invented nonsensicalities\nwhile AI\nis by no means funny and\nAI researchers are convinced\nthey talk reasonably\nwhen reading the AI literature, we are confronted with all kinds of absurdities, like machines\nhaving\nmotives\nwesternized\ndesires\nor\nhuman-like bodies, or\nself replicating intelligent machines\nthe engineering perspective of the field largely underestimates our deep ignorance about the biological\nbasis\nof those processes\n“if you want to mimic something you should know what it is\nthe man-a-machine\nis not just\nsomehow problematic\nbut\nit is\nabsurd\nridiculousy nonsensical",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process. | we must ask in what way
the human must necessarily be different from
the machine. The answer is
so obvious
it makes the very asking of the question appear ludicrous, if
not obscene. It is a sign of
madness
that this
is to be addressed at all
Weizenbaum
probably thought
as science progressed, the answer would
become obvious to everyone in the field
this did not happen. The opposite happened
Today
it is common “knowledge”
that computers will, sooner or later, be human-like
So was Weizenbaum
proven wrong?
Quite
the contrary
the debate is largely dominated by engineers, mathematicians and info
scientists
psychologists
the specialists responsible for the
mental functions the computer scientists intend to emulate
contributed little to instructing those
scientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human
wave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or
superhuman, and that they eventually will take over control of the world
I want to refute the idea that there can be
machines comparable to
humans
I soon realized that my task would be comparable to Sisyphus’
With the latest
accomplishment popping up
weekly
I would have to
push
the rock
every
week with no end in sight
The book would have to be rewritten
once a year. Machines with human-like
functions
have been predicted for
decades
this will go on for
decades more unless the scientific community realizes
there are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons
to make no sense whatsoever
A comparison between
human and
computer is not seen as mad in the AI community
not because obvious similarities could be shown to exist
there aren’t any
but
because this kind of speak
has become
common among AI researchers. Take a typical phrase like “culture is poured into artificial brains”
You can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely
may not be felt as nonsense any more. The
phrase about how culture can
get into a machine reminds one in its nonsensicality of
The head is round
to allow
thoughts to change direction
much of AI speak appears
inspired by Dadaism. The only difference between AI and Dada
is that Dada is funny and the Dadaists consciously invented nonsensicalities
while AI
is by no means funny and
AI researchers are convinced
they talk reasonably
when reading the AI literature, we are confronted with all kinds of absurdities, like machines
having
motives
westernized
desires
or
human-like bodies, or
self replicating intelligent machines
the engineering perspective of the field largely underestimates our deep ignorance about the biological
basis
of those processes
“if you want to mimic something you should know what it is
the man-a-machine
is not just
somehow problematic
but
it is
absurd
ridiculousy nonsensical | must
different
so obvious
very asking of the question
ludicrous
obscene
madness
addressed at all
did not happen
opposite happened
“knowledge”
sooner or later
human-like
proven wrong?
contrary
engineers
mathematicians
info
scientists
psychologists
specialists
intend to emulate
contribute
little
instructing
nature of human mental functions
grave definitional problems that plague the science of psychology
ride
computers-can-be-human
wave
furthered
criticized
outlandish ideas
superintelligent
superhuman
take over control of the world
refute
comparable to Sisyphus’
latest
accomplishment
weekly
push
the rock
every
week
no end in sight
rewritten
once a year
decades
go on
decades more
realizes
qualitative differences
human mental functions
functions running on computers
quantitative comparisons
no sense whatsoever
not
AI community
not
shown to exist
aren’t any
kind of speak
common among AI researchers
typical
poured into artificial brains
as little as you can pour the theory of relativity into a teacup
Any nonsense can be brought forward in a gramatically correct form
routinely
not be felt as nonsense any more
into a machine
nonsensicality
round
thoughts
change direction
inspired by Dadaism
difference
Dada is funny
consciously
AI
no means
AI researchers
convinced
they talk reasonably
all kinds of absurdities
machine
motives
desires
human-like bodies
self replicating intelligent
engineering
underestimates
deep ignorance
biological
basis
know what it is
just
problematic
absurd
ridiculousy nonsensical | 219 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process.",
"role": "system"
},
{
"content": "we must ask in what way\nthe human must necessarily be different from\nthe machine. The answer is\nso obvious\nit makes the very asking of the question appear ludicrous, if\nnot obscene. It is a sign of\nmadness\nthat this\nis to be addressed at all\nWeizenbaum\nprobably thought\nas science progressed, the answer would\nbecome obvious to everyone in the field\nthis did not happen. The opposite happened\nToday\nit is common “knowledge”\nthat computers will, sooner or later, be human-like\nSo was Weizenbaum\nproven wrong?\nQuite\nthe contrary\nthe debate is largely dominated by engineers, mathematicians and info\nscientists\npsychologists\nthe specialists responsible for the\nmental functions the computer scientists intend to emulate\ncontributed little to instructing those\nscientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human\nwave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or\nsuperhuman, and that they eventually will take over control of the world\nI want to refute the idea that there can be\nmachines comparable to\nhumans\nI soon realized that my task would be comparable to Sisyphus’\nWith the latest\naccomplishment popping up\nweekly\nI would have to\npush\nthe rock\nevery\nweek with no end in sight\nThe book would have to be rewritten\nonce a year. Machines with human-like\nfunctions\nhave been predicted for\ndecades\nthis will go on for\ndecades more unless the scientific community realizes\nthere are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons\nto make no sense whatsoever\nA comparison between\nhuman and\ncomputer is not seen as mad in the AI community\nnot because obvious similarities could be shown to exist\nthere aren’t any\nbut\nbecause this kind of speak\nhas become\ncommon among AI researchers. Take a typical phrase like “culture is poured into artificial brains”\nYou can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely\nmay not be felt as nonsense any more. The\nphrase about how culture can\nget into a machine reminds one in its nonsensicality of\nThe head is round\nto allow\nthoughts to change direction\nmuch of AI speak appears\ninspired by Dadaism. The only difference between AI and Dada\nis that Dada is funny and the Dadaists consciously invented nonsensicalities\nwhile AI\nis by no means funny and\nAI researchers are convinced\nthey talk reasonably\nwhen reading the AI literature, we are confronted with all kinds of absurdities, like machines\nhaving\nmotives\nwesternized\ndesires\nor\nhuman-like bodies, or\nself replicating intelligent machines\nthe engineering perspective of the field largely underestimates our deep ignorance about the biological\nbasis\nof those processes\n“if you want to mimic something you should know what it is\nthe man-a-machine\nis not just\nsomehow problematic\nbut\nit is\nabsurd\nridiculousy nonsensical",
"role": "user"
},
{
"content": "must\ndifferent\nso obvious\nvery asking of the question\nludicrous\nobscene\nmadness\naddressed at all\ndid not happen\nopposite happened\n“knowledge”\nsooner or later\nhuman-like\nproven wrong?\ncontrary\nengineers\nmathematicians\ninfo\nscientists\npsychologists\nspecialists\nintend to emulate\ncontribute\nlittle\ninstructing\nnature of human mental functions\ngrave definitional problems that plague the science of psychology\nride\ncomputers-can-be-human\nwave\nfurthered\ncriticized\noutlandish ideas\nsuperintelligent\nsuperhuman\ntake over control of the world\nrefute\ncomparable to Sisyphus’\nlatest\naccomplishment\nweekly\npush\nthe rock\nevery\nweek\nno end in sight\nrewritten\nonce a year\ndecades\ngo on\ndecades more\nrealizes\nqualitative differences\nhuman mental functions\nfunctions running on computers\nquantitative comparisons\nno sense whatsoever\nnot\nAI community\nnot\nshown to exist\naren’t any\nkind of speak\ncommon among AI researchers\ntypical\npoured into artificial brains\nas little as you can pour the theory of relativity into a teacup\nAny nonsense can be brought forward in a gramatically correct form\nroutinely\nnot be felt as nonsense any more\ninto a machine\nnonsensicality\nround\nthoughts\nchange direction\ninspired by Dadaism\ndifference\nDada is funny\nconsciously\nAI\nno means\nAI researchers\nconvinced\nthey talk reasonably\nall kinds of absurdities\nmachine\nmotives\ndesires\nhuman-like bodies\nself replicating intelligent\nengineering\nunderestimates\ndeep ignorance\nbiological\nbasis\nknow what it is\njust\nproblematic\nabsurd\nridiculousy nonsensical",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process.
Highlight Ratio: 0.2 | we must ask in what way
the human must necessarily be different from
the machine. The answer is
so obvious
it makes the very asking of the question appear ludicrous, if
not obscene. It is a sign of
madness
that this
is to be addressed at all
Weizenbaum
probably thought
as science progressed, the answer would
become obvious to everyone in the field
this did not happen. The opposite happened
Today
it is common “knowledge”
that computers will, sooner or later, be human-like
So was Weizenbaum
proven wrong?
Quite
the contrary
the debate is largely dominated by engineers, mathematicians and info
scientists
psychologists
the specialists responsible for the
mental functions the computer scientists intend to emulate
contributed little to instructing those
scientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human
wave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or
superhuman, and that they eventually will take over control of the world
I want to refute the idea that there can be
machines comparable to
humans
I soon realized that my task would be comparable to Sisyphus’
With the latest
accomplishment popping up
weekly
I would have to
push
the rock
every
week with no end in sight
The book would have to be rewritten
once a year. Machines with human-like
functions
have been predicted for
decades
this will go on for
decades more unless the scientific community realizes
there are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons
to make no sense whatsoever
A comparison between
human and
computer is not seen as mad in the AI community
not because obvious similarities could be shown to exist
there aren’t any
but
because this kind of speak
has become
common among AI researchers. Take a typical phrase like “culture is poured into artificial brains”
You can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely
may not be felt as nonsense any more. The
phrase about how culture can
get into a machine reminds one in its nonsensicality of
The head is round
to allow
thoughts to change direction
much of AI speak appears
inspired by Dadaism. The only difference between AI and Dada
is that Dada is funny and the Dadaists consciously invented nonsensicalities
while AI
is by no means funny and
AI researchers are convinced
they talk reasonably
when reading the AI literature, we are confronted with all kinds of absurdities, like machines
having
motives
westernized
desires
or
human-like bodies, or
self replicating intelligent machines
the engineering perspective of the field largely underestimates our deep ignorance about the biological
basis
of those processes
“if you want to mimic something you should know what it is
the man-a-machine
is not just
somehow problematic
but
it is
absurd
ridiculousy nonsensical | the human must
be different
so obvious
it makes the very asking
obscene
debate is
dominated by engineers
psychologists
contribute
little
They
furthered
such outlandish ideas as
computers becoming superintelligent
to refute
would be
Sisyphus’
With the latest
accomplishment
weekly
unless the
community realizes
there are qualitative differences
which cause
comparisons
make no sense whatsoever
a typical phrase like “culture
poured into artificial brains
reminds
in
nonsensicality of
Dadaism. The only difference
is
Dada is funny and
AI researchers are convinced
they talk reasonably
absurdities, like machine
motives
or
self replicating
underestimates
deep ignorance about
biological
basis | 220 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: It’s nonsense produced by AI industry groupthink that ignores basic knowledge about cognitive process.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "we must ask in what way\nthe human must necessarily be different from\nthe machine. The answer is\nso obvious\nit makes the very asking of the question appear ludicrous, if\nnot obscene. It is a sign of\nmadness\nthat this\nis to be addressed at all\nWeizenbaum\nprobably thought\nas science progressed, the answer would\nbecome obvious to everyone in the field\nthis did not happen. The opposite happened\nToday\nit is common “knowledge”\nthat computers will, sooner or later, be human-like\nSo was Weizenbaum\nproven wrong?\nQuite\nthe contrary\nthe debate is largely dominated by engineers, mathematicians and info\nscientists\npsychologists\nthe specialists responsible for the\nmental functions the computer scientists intend to emulate\ncontributed little to instructing those\nscientists about the nature of human mental functions or the grave definitional problems that plague the science of psychology. They rather decided to ride the computers-can-be-human\nwave and thus furthered rather than criticized that flawed project, promulgating even such outlandish ideas as that of computers becoming superintelligent or\nsuperhuman, and that they eventually will take over control of the world\nI want to refute the idea that there can be\nmachines comparable to\nhumans\nI soon realized that my task would be comparable to Sisyphus’\nWith the latest\naccomplishment popping up\nweekly\nI would have to\npush\nthe rock\nevery\nweek with no end in sight\nThe book would have to be rewritten\nonce a year. Machines with human-like\nfunctions\nhave been predicted for\ndecades\nthis will go on for\ndecades more unless the scientific community realizes\nthere are qualitative differences between human mental functions and functions running on computers, which cause quantitative comparisons\nto make no sense whatsoever\nA comparison between\nhuman and\ncomputer is not seen as mad in the AI community\nnot because obvious similarities could be shown to exist\nthere aren’t any\nbut\nbecause this kind of speak\nhas become\ncommon among AI researchers. Take a typical phrase like “culture is poured into artificial brains”\nYou can pour culture into computers as little as you can pour the theory of relativity into a teacup. Any nonsense can be brought forward in a gramatically correct form and, if done routinely\nmay not be felt as nonsense any more. The\nphrase about how culture can\nget into a machine reminds one in its nonsensicality of\nThe head is round\nto allow\nthoughts to change direction\nmuch of AI speak appears\ninspired by Dadaism. The only difference between AI and Dada\nis that Dada is funny and the Dadaists consciously invented nonsensicalities\nwhile AI\nis by no means funny and\nAI researchers are convinced\nthey talk reasonably\nwhen reading the AI literature, we are confronted with all kinds of absurdities, like machines\nhaving\nmotives\nwesternized\ndesires\nor\nhuman-like bodies, or\nself replicating intelligent machines\nthe engineering perspective of the field largely underestimates our deep ignorance about the biological\nbasis\nof those processes\n“if you want to mimic something you should know what it is\nthe man-a-machine\nis not just\nsomehow problematic\nbut\nit is\nabsurd\nridiculousy nonsensical",
"role": "user"
},
{
"content": "the human must\nbe different\nso obvious\nit makes the very asking\nobscene\ndebate is\ndominated by engineers\npsychologists\ncontribute\nlittle\nThey\nfurthered\nsuch outlandish ideas as\ncomputers becoming superintelligent\nto refute\nwould be\nSisyphus’\nWith the latest\naccomplishment\nweekly\nunless the\ncommunity realizes\nthere are qualitative differences\nwhich cause\ncomparisons\nmake no sense whatsoever\na typical phrase like “culture\npoured into artificial brains\nreminds\nin\nnonsensicality of\nDadaism. The only difference\nis\nDada is funny and\nAI researchers are convinced\nthey talk reasonably\nabsurdities, like machine\nmotives\nor\nself replicating\nunderestimates\ndeep ignorance about\nbiological\nbasis",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning. | 9.2 The limitations of deep learning
The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach for current deeplearning techniques—even given vast amounts of human-annotated data. Say, for instance, that you could assemble a dataset of hundreds of thousands—even millions—of English-language descriptions of the features of a software product, written by a product manager, as well as the corresponding source code developed by a team of engineers to meet these requirements. Even with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase. That’s just one example among many. In general, anything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw at them. Even learning a sorting algorithm with a deep neural network is tremendously difficult.
This is because a deep-learning model is just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform from X to Y. A deep-learning model can be interpreted as a kind of program; but, inversely, most programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves the task or, even if one exists, it may not be learnable: the corresponding geometric transform may be far too complex, or there may not be appropriate data available to learn it.
Scaling up current deep-learning techniques by stacking more layers and using more training data can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent and that most of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold.
9.2.1 The risk of anthropomorphizing machine-learning models
One real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions, beliefs, and knowledge on the things around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means that, for instance, when we’re able to somewhat successfully train a model to generate captions to describe pictures, we’re led to believe that the model “understands” the contents of the pictures and the captions it generates. Then we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions (see figure 9.1).
[FIGURE 9.1 OMITTED]
In particular, this is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model into misclassifying them. You’re already aware that, for instance, it’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some convnet filter—this is the basis of the filter-visualization technique introduced in chapter 5, as well as the DeepDream algorithm in chapter 8. Similarly, through gradient ascent, you can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding to it a gibbon gradient, we can get a neural network to classify the panda as a gibbon (see figure 9.2). This evidences both the brittleness of these models and the deep difference between their input-to-output mapping and our human perception.
In short, deep-learning models don’t have any understanding of their input—at least, not in a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds—the one developed from our experience as embodied agents. It’s like a dim image in a mirror (see figure 9.3).
[FIGURE 9.3 OMITTED]
As a machine-learning practitioner, always be mindful of this, and never fall into the trap of believing that neural networks understand the task they perform—they don’t, at least not in a way that would make sense to us. They were trained on a different, far narrower task than the one we wanted to teach them: that of mapping training inputs to training targets, point by point. Show them anything that deviates from their training data, and they will break in absurd ways.
9.2.2 Local generalization vs. extreme generalization
There are fundamental differences between the straightforward geometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience instead of being presented with explicit training examples. In addition to the different learning processes, there’s a basic difference in the nature of the underlying representations.
Humans are capable of far more than mapping immediate stimuli to immediate responses, as a deep network, or maybe an insect, would. We maintain complex, abstract models of our current situation, of ourselves, and of other people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge together known concepts to represent something we’ve never experienced before—like picturing a horse wearing jeans, for instance, or imagining what we’d do if we won the lottery. This ability to handle hypotheticals, to expand our mental model space far beyond what we can experience directly—to perform abstraction and reasoning—is arguably the defining characteristic of human cognition. I call it extreme generalization: an ability to adapt to novel, never-before-experienced situations using little data or even no new data at all.
This stands in sharp contrast with what deep nets do, which I call local generalization (see figure 9.4). The mapping from inputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time. Consider, for instance, the problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping from input space to output space. In contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials. Similarly, if you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city without getting hit by cars, the net would have to die many thousands of times in various situations until it could infer that cars are dangerous, and develop appropriate avoidance behaviors. Dropped into a new city, the net would have to relearn most of what it knows. On the other hand, humans are able to learn safe behaviors without having to die even once—again, thanks to our power of abstract modeling of hypothetical situations.
[FIGURE 9.4 OMITTED]
In short, despite our progress on machine perception, we’re still far from human-level AI. Our models can only perform local generalization, adapting to new situations that must be similar to past data, whereas human cognition is capable of extreme generalization, quickly adapting to radically novel situations and planning for long-term future situations.
9.2.3 Wrapping up
Here’s what you should remember: the only real success of deep learning so far has been the ability to map space X to space Y using a continuous geometric transform, given large amounts of human-annotated data. Doing this well is a game-changer for essentially every industry, but it’s still a long way from human-level AI.
To lift some of the limitations we have discussed and create AI that can compete with human brains, we need to move away from straightforward input-to-output mappings and on to reasoning and abstraction. A likely appropriate substrate for abstract modeling of various situations and concepts is that of computer programs. We said previously that machine-learning models can be defined as learnable programs; currently we can only learn programs that belong to a narrow and specific subset of all possible programs. But what if we could learn any program, in a modular and reusable way? Let’s see in the next section what the road ahead may look like.
9.3 The future of deep learning
This is a more speculative section aimed at opening horizons for people who want to join a research program or begin doing independent research. Given what we know of how deep nets work, their limitations, and the current state of the research landscape, can we predict where things are headed in the medium term? Following are some purely personal thoughts. Note that I don’t have a crystal ball, so a lot of what I anticipate may fail to become reality. I’m sharing these predictions not because I expect them to be proven completely right in the future, but because they’re interesting and actionable in the present.
At a high level, these are the main directions in which I see promise:
Models closer to general-purpose computer programs, built on top of far richer primitives than the current differentiable layers. This is how we’ll get to reasoning and abstraction, the lack of which is the fundamental weakness of current models.
New forms of learning that make the previous point possible, allowing models to move away from differentiable transforms.
Models that require less involvement from human engineers. It shouldn’t be your job to tune knobs endlessly.
Greater, systematic reuse of previously learned features and architectures, such as metalearning systems using reusable and modular program subroutines.
Additionally, note that these considerations aren’t specific to the sort of supervised learning that has been the bread and butter of deep learning so far—rather, they’re applicable to any form of machine learning, including unsupervised, self-supervised, and reinforcement learning. It isn’t fundamentally important where your labels come from or what your training loop looks like; these different branches of machine learning are different facets of the same construct. Let’s dive in.
9.3.1 Models as programs
As noted in the previous section, a necessary transformational development that we can expect in the field of machine learning is a move away from models that perform purely pattern recognition and can only achieve local generalization, toward models capable of abstraction and reasoning that can achieve extreme generalization. Current AI programs that are capable of basic forms of reasoning are all hardcoded by human programmers: for instance, software that relies on search algorithms, graph manipulation, and formal logic. In DeepMind’s AlphaGo, for example, most of the intelligence on display is designed and hardcoded by expert programmers (such as Monte Carlo Tree Search); learning from data happens only in specialized submodules (value networks and policy networks). But in the future, such AI systems may be fully learned, with no human involvement.
What path could make this happen? Consider a well-known type of network: RNNs. It’s important to note that RNNs have slightly fewer limitations than feedforward networks. That’s because RNNs are a bit more than mere geometric transformations: they’re geometric transformations repeatedly applied inside a for loop. The temporal for loop is itself hardcoded by human developers: it’s a built-in assumption of the network. Naturally, RNNs are still extremely limited in what they can represent, primarily because each step they perform is a differentiable geometric transformation, and they carry information from step to step via points in a continuous geometric space (state vectors). Now imagine a neural network that’s augmented in a similar way with programming primitives—but instead of a single hardcoded for loop with hardcoded geometric memory, the network includes a large set of programming primitives that the model is free to manipulate to expand its processing function, such as if branches, while statements, variable creation, disk storage for long-term memory, sorting operators, advanced data structures (such as lists, graphs, and hash tables), and many more. The space of programs that such a network could represent would be far broader than what can be represented with current deep-learning models, and some of these programs could achieve superior generalization power.
We’ll move away from having, on one hand, hardcoded algorithmic intelligence (handcrafted software) and, on the other hand, learned geometric intelligence (deep learning). Instead, we’ll have a blend of formal algorithmic modules that provide reasoning and abstraction capabilities, and geometric modules that provide informal intuition and pattern-recognition capabilities. The entire system will be learned with little or no human involvement.
A related subfield of AI that I think may be about to take off in a big way is program synthesis, in particular neural program synthesis. Program synthesis consists of automatically generating simple programs by using a search algorithm (possibly genetic search, as in genetic programming) to explore a large space of possible programs. The search stops when a program is found that matches the required specifications, often provided as a set of input-output pairs. This is highly reminiscent of machine learning: given training data provided as input-output pairs, we find a program that matches inputs to outputs and can generalize to new inputs. The difference is that instead of learning parameter values in a hardcoded program (a neural network), we generate source code via a discrete search process.
I definitely expect this subfield to see a wave of renewed interest in the next few years. In particular, I expect the emergence of a crossover subfield between deep learning and program synthesis, where instead of generating programs in a generalpurpose language, we’ll generate neural networks (geometric data-processing flows) augmented with a rich set of algorithmic primitives, such as for loops and many others (see figure 9.5). This should be far more tractable and useful than directly generating source code, and it will dramatically expand the scope of problems that can be solved with machine learning—the space of programs that we can generate automatically, given appropriate training data. Contemporary RNNs can be seen as a prehistoric ancestor of such hybrid algorithmic-geometric models.
[FIGURE 9.5 OMITTED]
9.3.2 Beyond backpropagation and differentiable layers
If machine-learning models become more like programs, then they will mostly no longer be differentiable—these programs will still use continuous geometric layers as subroutines, which will be differentiable, but the model as a whole won’t be. As a result, using backpropagation to adjust weight values in a fixed, hardcoded network can’t be the method of choice for training models in the future—at least, it can’t be the entire story. We need to figure out how to train non-differentiable systems efficiently. Current approaches include genetic algorithms, evolution strategies, certain reinforcement-learning methods, and alternating direction method of multipliers (ADMM). Naturally, gradient descent isn’t going anywhere; gradient information will always be useful for optimizing differentiable parametric functions. But our models will become increasingly more ambitious than mere differentiable parametric functions, and thus their automatic development (the learning in machine learning) will require more than backpropagation.
In addition, backpropagation is end to end, which is a great thing for learning good chained transformations but is computationally inefficient because it doesn’t fully take advantage of the modularity of deep networks. To make something more efficient, there’s one universal recipe: introduce modularity and hierarchy. So we can make backpropagation more efficient by introducing decoupled training modules with a synchronization mechanism between them, organized in a hierarchical fashion. This strategy is somewhat reflected in DeepMind’s recent work on synthetic gradients. I expect more along these lines in the near future. I can imagine a future where models that are globally non-differentiable (but feature differentiable parts) are trained— grown—using an efficient search process that doesn’t use gradients, whereas the differentiable parts are trained even faster by taking advantage of gradients using a more efficient version of backpropagation.
9.3.3 Automated machine learning
In the future, model architectures will be learned rather than be handcrafted by engineer-artisans. Learning architectures goes hand in hand with the use of richer sets of primitives and program-like machine-learning models.
Currently, most of the job of a deep-learning engineer consists of munging data with Python scripts and then tuning the architecture and hyperparameters of a deep network at length to get a working model—or even to get a state-of-the-art model, if the engineer is that ambitious. Needless to say, that isn’t an optimal setup. But AI can help. Unfortunately, the data-munging part is tough to automate, because it often requires domain knowledge as well as a clear, high-level understanding of what the engineer wants to achieve. Hyperparameter tuning, however, is a simple search procedure; and in that case we know what the engineer wants to achieve: it’s defined by the loss function of the network being tuned. It’s already common practice to set up basic AutoML systems that take care of most model knob tuning. I even set up my own, years ago, to win Kaggle competitions.
At the most basic level, such a system would tune the number of layers in a stack, their order, and the number of units or filters in each layer. This is commonly done with libraries such as Hyperopt, which we discussed in chapter 7. But we can also be far more ambitious and attempt to learn an appropriate architecture from scratch, with as few constraints as possible: for instance, via reinforcement learning or genetic algorithms.
Another important AutoML direction involves learning model architecture jointly with model weights. Because training a new model from scratch every time we try a slightly different architecture is tremendously inefficient, a truly powerful AutoML system would evolve architectures at the same time the features of the model were being tuned via backpropagation on the training data. Such approaches are beginning to emerge as I write these lines.
When this starts to happen, the jobs of machine-learning engineers won’t disappear—rather, engineers will move up the value-creation chain. They will begin to put much more effort into crafting complex loss functions that truly reflect business goals and understanding how their models impact the digital ecosystems in which they’re deployed (for example, the users who consume the model’s predictions and generate the model’s training data)—problems that only the largest companies can afford to consider at present.
9.3.4 Lifelong learning and modular subroutine reuse
If models become more complex and are built on top of richer algorithmic primitives, then this increased complexity will require higher reuse between tasks, rather than training a new model from scratch every time we have a new task or a new dataset. Many datasets don’t contain enough information for us to develop a new, complex model from scratch, and it will be necessary to use information from previously encountered datasets (much as you don’t learn English from scratch every time you open a new book—that would be impossible). Training models from scratch on every new task is also inefficient due to the large overlap between the current tasks and previously encountered tasks.
A remarkable observation has been made repeatedly in recent years: training the same model to do several loosely connected tasks at the same time results in a model that’s better at each task. For instance, training the same neural machine-translation model to perform both English-to-German translation and French-to-Italian translation will result in a model that’s better at each language pair. Similarly, training an image-classification model jointly with an image-segmentation model, sharing the same convolutional base, results in a model that’s better at both tasks. This is fairly intuitive: there’s always some information overlap between seemingly disconnected tasks, and a joint model has access to a greater amount of information about each individual task than a model trained on that specific task only.
Currently, when it comes to model reuse across tasks, we use pretrained weights for models that perform common functions, such as visual feature extraction. You saw this in action in chapter 5. In the future, I expect a generalized version of this to be commonplace: we’ll use not only previously learned features (submodel weights) but also model architectures and training procedures. As models become more like programs, we’ll begin to reuse program subroutines like the functions and classes found in human programming languages.
Think of the process of software development today: once an engineer solves a specific problem (HTTP queries in Python, for instance), they package it as an abstract, reusable library. Engineers who face a similar problem in the future will be able to search for existing libraries, download one, and use it in their own project. In a similar way, in the future, metalearning systems will be able to assemble new programs by sifting through a global library of high-level reusable blocks. When the system finds itself developing similar program subroutines for several different tasks, it can come up with an abstract, reusable version of the subroutine and store it in the global library (see figure 9.6). Such a process will implement abstraction: a necessary component for achieving extreme generalization. A subroutine that’s useful across different tasks and domains can be said to abstract some aspect of problem solving. This definition of abstraction is similar to the notion of abstraction in software engineering. These subroutines can be either geometric (deep-learning modules with pretrained representations) or algorithmic (closer to the libraries that contemporary software engineers manipulate).
[FIGURE 9.6 OMITTED]
9.3.5 The long-term vision
In short, here’s my long-term vision for machine learning:
Models will be more like programs and will have capabilities that go far beyond the continuous geometric transformations of the input data we currently work with. These programs will arguably be much closer to the abstract mental models that humans maintain about their surroundings and themselves, and they will be capable of stronger generalization due to their rich algorithmic nature.
In particular, models will blend algorithmic modules providing formal reasoning, search, and abstraction capabilities with geometric modules providing informal intuition and pattern-recognition capabilities. AlphaGo (a system that required a lot of manual software engineering and human-made design decisions) provides an early example of what such a blend of symbolic and geometric AI could look like.
Such models will be grown automatically rather than hardcoded by human engineers, using modular parts stored in a global library of reusable subroutines—a library evolved by learning high-performing models on thousands of previous tasks and datasets. As frequent problem-solving patterns are identified by the meta-learning system, they will be turned into reusable subroutines—much like functions and classes in software engineering—and added to the global library. This will achieve abstraction.
This global library and associated model-growing system will be able to achieve some form of human-like extreme generalization: given a new task or situation, the system will be able to assemble a new working model appropriate for the task using very little data, thanks to rich program-like primitives that generalize well, and extensive experience with similar tasks. In the same way, humans can quickly learn to play a complex new video game if they have experience with many previous games, because the models derived from this previous experience are abstract and program-like, rather than a basic mapping between stimuli and action.
As such, this perpetually learning model-growing system can be interpreted as an artificial general intelligence (AGI). But don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology. Such a critique, however, doesn’t belong in this book. | The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach
even given vast amounts of human-annotated data. Say
you could assemble a dataset of
millions—of
descriptions of the features of a software product
as well as the corresponding source code
Even with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase
anything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw
Even learning a sorting algorithm with a deep neural network is tremendously difficult
This is because
deep-learning
is just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform
A deep-learning model can be interpreted as a kind of program; but
most programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves
or, even if one exists, it may not be learnable
the corresponding geometric transform may be far too complex, or there may not be appropriate data available
Scaling up current
techniques by stacking more layers and using more
data can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent
most of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold
One real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions
and knowledge on
things around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means
when we’re able to somewhat successfully train a model to generate captions
we’re led to believe that the model “understands” the contents
Then we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions
this is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model
it’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some
filter
you can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding
a gibbon gradient, we can get a neural network to classify the panda as a gibbon
This evidences both the brittleness of these models and the deep difference between their
mapping and our
perception
deep-learning models don’t have any understanding of their input
in a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds
like a dim image in a mirror
never fall into the trap of believing that neural networks understand the task they perform—they don’t
They were trained on a different, far narrower task
that of mapping training inputs to training targets, point by point. Show them anything that deviates from their training
and they will break in absurd ways
There are fundamental differences between the
geometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience
there’s a basic difference in the nature of the underlying representations.
Humans are capable of far more than mapping immediate stimuli to immediate responses
We maintain complex, abstract models of our
situation
ourselves
and
other people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge
known concepts to represent something we’ve never experienced before
This ability to handle hypotheticals, to expand our mental model space
beyond what we can experience directly—to perform abstraction and reasoning—is
the defining characteristic of human cognition
an ability to adapt to novel
situations using
no new data at all
This stands in sharp contrast with
deep nets
mapping
inputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time
Consider
the problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping
In contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials
if you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city
the net would have to die
thousands of times
Dropped into a new city, the net would have to relearn
what it knows
On the other hand, humans are able to learn safe behaviors without having to die even once
thanks to our power of abstract modeling of hypothetical situations
we’re
far from human-level AI
Given what we know of how deep nets work, their limitations, and the current state of
research
can we predict where things are headed
?
long-term
Models will be more like programs and will have capabilities that go
beyond the continuous geometric transformations of the input data we currently work with. These programs will
be
closer to
abstract mental models
models will blend algorithmic modules
This global library and associated model-growing system will be able to achieve some form of
generalization
But don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology | 221 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning.",
"role": "system"
},
{
"content": "9.2 The limitations of deep learning\nThe space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach for current deeplearning techniques—even given vast amounts of human-annotated data. Say, for instance, that you could assemble a dataset of hundreds of thousands—even millions—of English-language descriptions of the features of a software product, written by a product manager, as well as the corresponding source code developed by a team of engineers to meet these requirements. Even with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase. That’s just one example among many. In general, anything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw at them. Even learning a sorting algorithm with a deep neural network is tremendously difficult.\nThis is because a deep-learning model is just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform from X to Y. A deep-learning model can be interpreted as a kind of program; but, inversely, most programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves the task or, even if one exists, it may not be learnable: the corresponding geometric transform may be far too complex, or there may not be appropriate data available to learn it.\nScaling up current deep-learning techniques by stacking more layers and using more training data can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent and that most of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold.\n9.2.1 The risk of anthropomorphizing machine-learning models\nOne real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions, beliefs, and knowledge on the things around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means that, for instance, when we’re able to somewhat successfully train a model to generate captions to describe pictures, we’re led to believe that the model “understands” the contents of the pictures and the captions it generates. Then we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions (see figure 9.1).\n[FIGURE 9.1 OMITTED]\nIn particular, this is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model into misclassifying them. You’re already aware that, for instance, it’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some convnet filter—this is the basis of the filter-visualization technique introduced in chapter 5, as well as the DeepDream algorithm in chapter 8. Similarly, through gradient ascent, you can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding to it a gibbon gradient, we can get a neural network to classify the panda as a gibbon (see figure 9.2). This evidences both the brittleness of these models and the deep difference between their input-to-output mapping and our human perception.\n\nIn short, deep-learning models don’t have any understanding of their input—at least, not in a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds—the one developed from our experience as embodied agents. It’s like a dim image in a mirror (see figure 9.3).\n[FIGURE 9.3 OMITTED]\nAs a machine-learning practitioner, always be mindful of this, and never fall into the trap of believing that neural networks understand the task they perform—they don’t, at least not in a way that would make sense to us. They were trained on a different, far narrower task than the one we wanted to teach them: that of mapping training inputs to training targets, point by point. Show them anything that deviates from their training data, and they will break in absurd ways.\n9.2.2 Local generalization vs. extreme generalization\nThere are fundamental differences between the straightforward geometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience instead of being presented with explicit training examples. In addition to the different learning processes, there’s a basic difference in the nature of the underlying representations.\nHumans are capable of far more than mapping immediate stimuli to immediate responses, as a deep network, or maybe an insect, would. We maintain complex, abstract models of our current situation, of ourselves, and of other people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge together known concepts to represent something we’ve never experienced before—like picturing a horse wearing jeans, for instance, or imagining what we’d do if we won the lottery. This ability to handle hypotheticals, to expand our mental model space far beyond what we can experience directly—to perform abstraction and reasoning—is arguably the defining characteristic of human cognition. I call it extreme generalization: an ability to adapt to novel, never-before-experienced situations using little data or even no new data at all.\nThis stands in sharp contrast with what deep nets do, which I call local generalization (see figure 9.4). The mapping from inputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time. Consider, for instance, the problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping from input space to output space. In contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials. Similarly, if you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city without getting hit by cars, the net would have to die many thousands of times in various situations until it could infer that cars are dangerous, and develop appropriate avoidance behaviors. Dropped into a new city, the net would have to relearn most of what it knows. On the other hand, humans are able to learn safe behaviors without having to die even once—again, thanks to our power of abstract modeling of hypothetical situations.\n[FIGURE 9.4 OMITTED]\nIn short, despite our progress on machine perception, we’re still far from human-level AI. Our models can only perform local generalization, adapting to new situations that must be similar to past data, whereas human cognition is capable of extreme generalization, quickly adapting to radically novel situations and planning for long-term future situations.\n9.2.3 Wrapping up\nHere’s what you should remember: the only real success of deep learning so far has been the ability to map space X to space Y using a continuous geometric transform, given large amounts of human-annotated data. Doing this well is a game-changer for essentially every industry, but it’s still a long way from human-level AI.\nTo lift some of the limitations we have discussed and create AI that can compete with human brains, we need to move away from straightforward input-to-output mappings and on to reasoning and abstraction. A likely appropriate substrate for abstract modeling of various situations and concepts is that of computer programs. We said previously that machine-learning models can be defined as learnable programs; currently we can only learn programs that belong to a narrow and specific subset of all possible programs. But what if we could learn any program, in a modular and reusable way? Let’s see in the next section what the road ahead may look like.\n9.3 The future of deep learning\nThis is a more speculative section aimed at opening horizons for people who want to join a research program or begin doing independent research. Given what we know of how deep nets work, their limitations, and the current state of the research landscape, can we predict where things are headed in the medium term? Following are some purely personal thoughts. Note that I don’t have a crystal ball, so a lot of what I anticipate may fail to become reality. I’m sharing these predictions not because I expect them to be proven completely right in the future, but because they’re interesting and actionable in the present.\nAt a high level, these are the main directions in which I see promise:\n Models closer to general-purpose computer programs, built on top of far richer primitives than the current differentiable layers. This is how we’ll get to reasoning and abstraction, the lack of which is the fundamental weakness of current models.\n New forms of learning that make the previous point possible, allowing models to move away from differentiable transforms.\n Models that require less involvement from human engineers. It shouldn’t be your job to tune knobs endlessly.\n Greater, systematic reuse of previously learned features and architectures, such as metalearning systems using reusable and modular program subroutines.\nAdditionally, note that these considerations aren’t specific to the sort of supervised learning that has been the bread and butter of deep learning so far—rather, they’re applicable to any form of machine learning, including unsupervised, self-supervised, and reinforcement learning. It isn’t fundamentally important where your labels come from or what your training loop looks like; these different branches of machine learning are different facets of the same construct. Let’s dive in.\n9.3.1 Models as programs\nAs noted in the previous section, a necessary transformational development that we can expect in the field of machine learning is a move away from models that perform purely pattern recognition and can only achieve local generalization, toward models capable of abstraction and reasoning that can achieve extreme generalization. Current AI programs that are capable of basic forms of reasoning are all hardcoded by human programmers: for instance, software that relies on search algorithms, graph manipulation, and formal logic. In DeepMind’s AlphaGo, for example, most of the intelligence on display is designed and hardcoded by expert programmers (such as Monte Carlo Tree Search); learning from data happens only in specialized submodules (value networks and policy networks). But in the future, such AI systems may be fully learned, with no human involvement.\nWhat path could make this happen? Consider a well-known type of network: RNNs. It’s important to note that RNNs have slightly fewer limitations than feedforward networks. That’s because RNNs are a bit more than mere geometric transformations: they’re geometric transformations repeatedly applied inside a for loop. The temporal for loop is itself hardcoded by human developers: it’s a built-in assumption of the network. Naturally, RNNs are still extremely limited in what they can represent, primarily because each step they perform is a differentiable geometric transformation, and they carry information from step to step via points in a continuous geometric space (state vectors). Now imagine a neural network that’s augmented in a similar way with programming primitives—but instead of a single hardcoded for loop with hardcoded geometric memory, the network includes a large set of programming primitives that the model is free to manipulate to expand its processing function, such as if branches, while statements, variable creation, disk storage for long-term memory, sorting operators, advanced data structures (such as lists, graphs, and hash tables), and many more. The space of programs that such a network could represent would be far broader than what can be represented with current deep-learning models, and some of these programs could achieve superior generalization power.\nWe’ll move away from having, on one hand, hardcoded algorithmic intelligence (handcrafted software) and, on the other hand, learned geometric intelligence (deep learning). Instead, we’ll have a blend of formal algorithmic modules that provide reasoning and abstraction capabilities, and geometric modules that provide informal intuition and pattern-recognition capabilities. The entire system will be learned with little or no human involvement.\nA related subfield of AI that I think may be about to take off in a big way is program synthesis, in particular neural program synthesis. Program synthesis consists of automatically generating simple programs by using a search algorithm (possibly genetic search, as in genetic programming) to explore a large space of possible programs. The search stops when a program is found that matches the required specifications, often provided as a set of input-output pairs. This is highly reminiscent of machine learning: given training data provided as input-output pairs, we find a program that matches inputs to outputs and can generalize to new inputs. The difference is that instead of learning parameter values in a hardcoded program (a neural network), we generate source code via a discrete search process.\nI definitely expect this subfield to see a wave of renewed interest in the next few years. In particular, I expect the emergence of a crossover subfield between deep learning and program synthesis, where instead of generating programs in a generalpurpose language, we’ll generate neural networks (geometric data-processing flows) augmented with a rich set of algorithmic primitives, such as for loops and many others (see figure 9.5). This should be far more tractable and useful than directly generating source code, and it will dramatically expand the scope of problems that can be solved with machine learning—the space of programs that we can generate automatically, given appropriate training data. Contemporary RNNs can be seen as a prehistoric ancestor of such hybrid algorithmic-geometric models.\n[FIGURE 9.5 OMITTED]\n9.3.2 Beyond backpropagation and differentiable layers\nIf machine-learning models become more like programs, then they will mostly no longer be differentiable—these programs will still use continuous geometric layers as subroutines, which will be differentiable, but the model as a whole won’t be. As a result, using backpropagation to adjust weight values in a fixed, hardcoded network can’t be the method of choice for training models in the future—at least, it can’t be the entire story. We need to figure out how to train non-differentiable systems efficiently. Current approaches include genetic algorithms, evolution strategies, certain reinforcement-learning methods, and alternating direction method of multipliers (ADMM). Naturally, gradient descent isn’t going anywhere; gradient information will always be useful for optimizing differentiable parametric functions. But our models will become increasingly more ambitious than mere differentiable parametric functions, and thus their automatic development (the learning in machine learning) will require more than backpropagation.\nIn addition, backpropagation is end to end, which is a great thing for learning good chained transformations but is computationally inefficient because it doesn’t fully take advantage of the modularity of deep networks. To make something more efficient, there’s one universal recipe: introduce modularity and hierarchy. So we can make backpropagation more efficient by introducing decoupled training modules with a synchronization mechanism between them, organized in a hierarchical fashion. This strategy is somewhat reflected in DeepMind’s recent work on synthetic gradients. I expect more along these lines in the near future. I can imagine a future where models that are globally non-differentiable (but feature differentiable parts) are trained— grown—using an efficient search process that doesn’t use gradients, whereas the differentiable parts are trained even faster by taking advantage of gradients using a more efficient version of backpropagation.\n9.3.3 Automated machine learning\nIn the future, model architectures will be learned rather than be handcrafted by engineer-artisans. Learning architectures goes hand in hand with the use of richer sets of primitives and program-like machine-learning models.\nCurrently, most of the job of a deep-learning engineer consists of munging data with Python scripts and then tuning the architecture and hyperparameters of a deep network at length to get a working model—or even to get a state-of-the-art model, if the engineer is that ambitious. Needless to say, that isn’t an optimal setup. But AI can help. Unfortunately, the data-munging part is tough to automate, because it often requires domain knowledge as well as a clear, high-level understanding of what the engineer wants to achieve. Hyperparameter tuning, however, is a simple search procedure; and in that case we know what the engineer wants to achieve: it’s defined by the loss function of the network being tuned. It’s already common practice to set up basic AutoML systems that take care of most model knob tuning. I even set up my own, years ago, to win Kaggle competitions.\nAt the most basic level, such a system would tune the number of layers in a stack, their order, and the number of units or filters in each layer. This is commonly done with libraries such as Hyperopt, which we discussed in chapter 7. But we can also be far more ambitious and attempt to learn an appropriate architecture from scratch, with as few constraints as possible: for instance, via reinforcement learning or genetic algorithms.\nAnother important AutoML direction involves learning model architecture jointly with model weights. Because training a new model from scratch every time we try a slightly different architecture is tremendously inefficient, a truly powerful AutoML system would evolve architectures at the same time the features of the model were being tuned via backpropagation on the training data. Such approaches are beginning to emerge as I write these lines.\nWhen this starts to happen, the jobs of machine-learning engineers won’t disappear—rather, engineers will move up the value-creation chain. They will begin to put much more effort into crafting complex loss functions that truly reflect business goals and understanding how their models impact the digital ecosystems in which they’re deployed (for example, the users who consume the model’s predictions and generate the model’s training data)—problems that only the largest companies can afford to consider at present.\n9.3.4 Lifelong learning and modular subroutine reuse\nIf models become more complex and are built on top of richer algorithmic primitives, then this increased complexity will require higher reuse between tasks, rather than training a new model from scratch every time we have a new task or a new dataset. Many datasets don’t contain enough information for us to develop a new, complex model from scratch, and it will be necessary to use information from previously encountered datasets (much as you don’t learn English from scratch every time you open a new book—that would be impossible). Training models from scratch on every new task is also inefficient due to the large overlap between the current tasks and previously encountered tasks.\nA remarkable observation has been made repeatedly in recent years: training the same model to do several loosely connected tasks at the same time results in a model that’s better at each task. For instance, training the same neural machine-translation model to perform both English-to-German translation and French-to-Italian translation will result in a model that’s better at each language pair. Similarly, training an image-classification model jointly with an image-segmentation model, sharing the same convolutional base, results in a model that’s better at both tasks. This is fairly intuitive: there’s always some information overlap between seemingly disconnected tasks, and a joint model has access to a greater amount of information about each individual task than a model trained on that specific task only.\nCurrently, when it comes to model reuse across tasks, we use pretrained weights for models that perform common functions, such as visual feature extraction. You saw this in action in chapter 5. In the future, I expect a generalized version of this to be commonplace: we’ll use not only previously learned features (submodel weights) but also model architectures and training procedures. As models become more like programs, we’ll begin to reuse program subroutines like the functions and classes found in human programming languages.\nThink of the process of software development today: once an engineer solves a specific problem (HTTP queries in Python, for instance), they package it as an abstract, reusable library. Engineers who face a similar problem in the future will be able to search for existing libraries, download one, and use it in their own project. In a similar way, in the future, metalearning systems will be able to assemble new programs by sifting through a global library of high-level reusable blocks. When the system finds itself developing similar program subroutines for several different tasks, it can come up with an abstract, reusable version of the subroutine and store it in the global library (see figure 9.6). Such a process will implement abstraction: a necessary component for achieving extreme generalization. A subroutine that’s useful across different tasks and domains can be said to abstract some aspect of problem solving. This definition of abstraction is similar to the notion of abstraction in software engineering. These subroutines can be either geometric (deep-learning modules with pretrained representations) or algorithmic (closer to the libraries that contemporary software engineers manipulate).\n[FIGURE 9.6 OMITTED]\n9.3.5 The long-term vision\nIn short, here’s my long-term vision for machine learning:\n Models will be more like programs and will have capabilities that go far beyond the continuous geometric transformations of the input data we currently work with. These programs will arguably be much closer to the abstract mental models that humans maintain about their surroundings and themselves, and they will be capable of stronger generalization due to their rich algorithmic nature.\n In particular, models will blend algorithmic modules providing formal reasoning, search, and abstraction capabilities with geometric modules providing informal intuition and pattern-recognition capabilities. AlphaGo (a system that required a lot of manual software engineering and human-made design decisions) provides an early example of what such a blend of symbolic and geometric AI could look like.\n Such models will be grown automatically rather than hardcoded by human engineers, using modular parts stored in a global library of reusable subroutines—a library evolved by learning high-performing models on thousands of previous tasks and datasets. As frequent problem-solving patterns are identified by the meta-learning system, they will be turned into reusable subroutines—much like functions and classes in software engineering—and added to the global library. This will achieve abstraction.\n This global library and associated model-growing system will be able to achieve some form of human-like extreme generalization: given a new task or situation, the system will be able to assemble a new working model appropriate for the task using very little data, thanks to rich program-like primitives that generalize well, and extensive experience with similar tasks. In the same way, humans can quickly learn to play a complex new video game if they have experience with many previous games, because the models derived from this previous experience are abstract and program-like, rather than a basic mapping between stimuli and action.\n As such, this perpetually learning model-growing system can be interpreted as an artificial general intelligence (AGI). But don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology. Such a critique, however, doesn’t belong in this book.",
"role": "user"
},
{
"content": "The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach\neven given vast amounts of human-annotated data. Say\nyou could assemble a dataset of\nmillions—of\ndescriptions of the features of a software product\nas well as the corresponding source code\nEven with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase\nanything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw\nEven learning a sorting algorithm with a deep neural network is tremendously difficult\nThis is because\ndeep-learning\nis just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform\nA deep-learning model can be interpreted as a kind of program; but\nmost programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves\nor, even if one exists, it may not be learnable\nthe corresponding geometric transform may be far too complex, or there may not be appropriate data available\nScaling up current\ntechniques by stacking more layers and using more\ndata can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent\nmost of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold\nOne real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions\nand knowledge on\nthings around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means\nwhen we’re able to somewhat successfully train a model to generate captions\nwe’re led to believe that the model “understands” the contents\nThen we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions\nthis is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model\nit’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some\nfilter\nyou can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding\na gibbon gradient, we can get a neural network to classify the panda as a gibbon\nThis evidences both the brittleness of these models and the deep difference between their\nmapping and our\nperception\ndeep-learning models don’t have any understanding of their input\nin a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds\nlike a dim image in a mirror\nnever fall into the trap of believing that neural networks understand the task they perform—they don’t\nThey were trained on a different, far narrower task\nthat of mapping training inputs to training targets, point by point. Show them anything that deviates from their training\nand they will break in absurd ways\nThere are fundamental differences between the\ngeometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience\nthere’s a basic difference in the nature of the underlying representations.\nHumans are capable of far more than mapping immediate stimuli to immediate responses\nWe maintain complex, abstract models of our\nsituation\nourselves\nand\nother people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge\nknown concepts to represent something we’ve never experienced before\nThis ability to handle hypotheticals, to expand our mental model space\nbeyond what we can experience directly—to perform abstraction and reasoning—is\nthe defining characteristic of human cognition\nan ability to adapt to novel\nsituations using\nno new data at all\nThis stands in sharp contrast with\ndeep nets\nmapping\ninputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time\nConsider\nthe problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping\nIn contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials\nif you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city\nthe net would have to die\nthousands of times\nDropped into a new city, the net would have to relearn\nwhat it knows\nOn the other hand, humans are able to learn safe behaviors without having to die even once\nthanks to our power of abstract modeling of hypothetical situations\nwe’re\nfar from human-level AI\nGiven what we know of how deep nets work, their limitations, and the current state of\nresearch\ncan we predict where things are headed\n?\nlong-term\nModels will be more like programs and will have capabilities that go\nbeyond the continuous geometric transformations of the input data we currently work with. These programs will\nbe\ncloser to\nabstract mental models\nmodels will blend algorithmic modules\nThis global library and associated model-growing system will be able to achieve some form of\ngeneralization\nBut don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning. | The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach
even given vast amounts of human-annotated data. Say
you could assemble a dataset of
millions—of
descriptions of the features of a software product
as well as the corresponding source code
Even with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase
anything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw
Even learning a sorting algorithm with a deep neural network is tremendously difficult
This is because
deep-learning
is just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform
A deep-learning model can be interpreted as a kind of program; but
most programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves
or, even if one exists, it may not be learnable
the corresponding geometric transform may be far too complex, or there may not be appropriate data available
Scaling up current
techniques by stacking more layers and using more
data can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent
most of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold
One real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions
and knowledge on
things around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means
when we’re able to somewhat successfully train a model to generate captions
we’re led to believe that the model “understands” the contents
Then we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions
this is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model
it’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some
filter
you can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding
a gibbon gradient, we can get a neural network to classify the panda as a gibbon
This evidences both the brittleness of these models and the deep difference between their
mapping and our
perception
deep-learning models don’t have any understanding of their input
in a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds
like a dim image in a mirror
never fall into the trap of believing that neural networks understand the task they perform—they don’t
They were trained on a different, far narrower task
that of mapping training inputs to training targets, point by point. Show them anything that deviates from their training
and they will break in absurd ways
There are fundamental differences between the
geometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience
there’s a basic difference in the nature of the underlying representations.
Humans are capable of far more than mapping immediate stimuli to immediate responses
We maintain complex, abstract models of our
situation
ourselves
and
other people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge
known concepts to represent something we’ve never experienced before
This ability to handle hypotheticals, to expand our mental model space
beyond what we can experience directly—to perform abstraction and reasoning—is
the defining characteristic of human cognition
an ability to adapt to novel
situations using
no new data at all
This stands in sharp contrast with
deep nets
mapping
inputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time
Consider
the problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping
In contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials
if you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city
the net would have to die
thousands of times
Dropped into a new city, the net would have to relearn
what it knows
On the other hand, humans are able to learn safe behaviors without having to die even once
thanks to our power of abstract modeling of hypothetical situations
we’re
far from human-level AI
Given what we know of how deep nets work, their limitations, and the current state of
research
can we predict where things are headed
?
long-term
Models will be more like programs and will have capabilities that go
beyond the continuous geometric transformations of the input data we currently work with. These programs will
be
closer to
abstract mental models
models will blend algorithmic modules
This global library and associated model-growing system will be able to achieve some form of
generalization
But don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology | applications
near
infinite
many
completely out of reach
vast amounts
millions
features
corresponding source code
this data
not train a deep-learning model to read a product description and generate the appropriate codebase
reasoning
programming
scientific method
long-term planning
out of reach
how much data
sorting
tremendously difficult
chain
simple, continuous geometric transformations mapping one vector space into another
map one data manifold X into another manifold Y
learnable continuous transform
most programs can’t be expressed as deep-learning models
either
exists no corresponding deep-neural network
solves
exists
not be learnable
far too complex
be
data available
Scaling up current
stacking more layers
more
data
superficially
some
fundamental
limited in what they can represent
most
programs
wish
can’t be expressed
continuous geometric morphing of a data manifold
misinterpreting what deep-learning models do
overestimating their abilities
humans
theory of mind
tend
project intent
knowledge
things around us
a smiley face on a rock
in our minds
we
somewhat successfully train
generate captions
we
believe
“understands”
contents
surprised
slight departure
completely absurd captions
adversarial examples
fed
designed to trick the model
maximize
activation of some
filter
modify
maximize the class prediction for a given class
gibbon gradient
get a neural network to classify the panda as a gibbon
brittleness
deep difference
their
mapping
our
perception
any understanding of their input
in a human sense
sensorimotor experience as humans
no access
experiences
can’t understand
human-relatable way
learn a geometric transform
maps data
simplistic sketch
original model in our minds
dim image in a mirror
understand
don’t
narrower
mapping training inputs to training targets, point by point
deviates
break in absurd ways
fundamental differences
geometric morphing from input to output
deep-learning models
humans
only
themselves
embodied experience
nature of the underlying representations
far more
immediate stimuli
immediate responses
complex, abstract models
situation
ourselves
other people
anticipate different possible futures
long-term planning
merge
never experienced before
hypotheticals
expand
beyond
experience directly
abstraction
reasoning
defining characteristic of human cognition
adapt
novel
situations
no new data at all
sharp contrast
deep nets
stops making sense
new inputs
even slightly
training time
dense sampling of the input space
reliable mapping
power of abstraction
physical models
derive
one or a few trials
human body
safely navigate a city
net
die
thousands of times
relearn
without having to die even once
far
predict
headed
?
long-term
more like programs
beyond
currently
closer
blend
form
generalization
don’t expect any singularitarian robot apocalypse to ensue
pure fantasy
long series of profound misunderstandings of both intelligence and tech | 222 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning.",
"role": "system"
},
{
"content": "The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach\neven given vast amounts of human-annotated data. Say\nyou could assemble a dataset of\nmillions—of\ndescriptions of the features of a software product\nas well as the corresponding source code\nEven with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase\nanything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw\nEven learning a sorting algorithm with a deep neural network is tremendously difficult\nThis is because\ndeep-learning\nis just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform\nA deep-learning model can be interpreted as a kind of program; but\nmost programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves\nor, even if one exists, it may not be learnable\nthe corresponding geometric transform may be far too complex, or there may not be appropriate data available\nScaling up current\ntechniques by stacking more layers and using more\ndata can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent\nmost of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold\nOne real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions\nand knowledge on\nthings around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means\nwhen we’re able to somewhat successfully train a model to generate captions\nwe’re led to believe that the model “understands” the contents\nThen we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions\nthis is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model\nit’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some\nfilter\nyou can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding\na gibbon gradient, we can get a neural network to classify the panda as a gibbon\nThis evidences both the brittleness of these models and the deep difference between their\nmapping and our\nperception\ndeep-learning models don’t have any understanding of their input\nin a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds\nlike a dim image in a mirror\nnever fall into the trap of believing that neural networks understand the task they perform—they don’t\nThey were trained on a different, far narrower task\nthat of mapping training inputs to training targets, point by point. Show them anything that deviates from their training\nand they will break in absurd ways\nThere are fundamental differences between the\ngeometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience\nthere’s a basic difference in the nature of the underlying representations.\nHumans are capable of far more than mapping immediate stimuli to immediate responses\nWe maintain complex, abstract models of our\nsituation\nourselves\nand\nother people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge\nknown concepts to represent something we’ve never experienced before\nThis ability to handle hypotheticals, to expand our mental model space\nbeyond what we can experience directly—to perform abstraction and reasoning—is\nthe defining characteristic of human cognition\nan ability to adapt to novel\nsituations using\nno new data at all\nThis stands in sharp contrast with\ndeep nets\nmapping\ninputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time\nConsider\nthe problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping\nIn contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials\nif you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city\nthe net would have to die\nthousands of times\nDropped into a new city, the net would have to relearn\nwhat it knows\nOn the other hand, humans are able to learn safe behaviors without having to die even once\nthanks to our power of abstract modeling of hypothetical situations\nwe’re\nfar from human-level AI\nGiven what we know of how deep nets work, their limitations, and the current state of\nresearch\ncan we predict where things are headed\n?\nlong-term\nModels will be more like programs and will have capabilities that go\nbeyond the continuous geometric transformations of the input data we currently work with. These programs will\nbe\ncloser to\nabstract mental models\nmodels will blend algorithmic modules\nThis global library and associated model-growing system will be able to achieve some form of\ngeneralization\nBut don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology",
"role": "user"
},
{
"content": "applications\nnear\ninfinite\nmany\ncompletely out of reach\nvast amounts\nmillions\nfeatures\ncorresponding source code\nthis data\nnot train a deep-learning model to read a product description and generate the appropriate codebase\nreasoning\nprogramming\nscientific method\nlong-term planning\nout of reach\nhow much data\nsorting\ntremendously difficult\nchain\nsimple, continuous geometric transformations mapping one vector space into another\nmap one data manifold X into another manifold Y\nlearnable continuous transform\nmost programs can’t be expressed as deep-learning models\neither\nexists no corresponding deep-neural network\nsolves\nexists\nnot be learnable\nfar too complex\nbe\ndata available\nScaling up current\nstacking more layers\nmore\ndata\nsuperficially\nsome\nfundamental\nlimited in what they can represent\nmost\nprograms\nwish\ncan’t be expressed\ncontinuous geometric morphing of a data manifold\nmisinterpreting what deep-learning models do\noverestimating their abilities\nhumans\ntheory of mind\ntend\nproject intent\nknowledge\nthings around us\na smiley face on a rock\nin our minds\nwe\nsomewhat successfully train\ngenerate captions\nwe\nbelieve\n“understands”\ncontents\nsurprised\nslight departure\ncompletely absurd captions\nadversarial examples\nfed\ndesigned to trick the model\nmaximize\nactivation of some\nfilter\nmodify\nmaximize the class prediction for a given class\ngibbon gradient\nget a neural network to classify the panda as a gibbon\nbrittleness\ndeep difference\ntheir\nmapping\nour\nperception\nany understanding of their input\nin a human sense\nsensorimotor experience as humans\nno access\nexperiences\ncan’t understand\nhuman-relatable way\nlearn a geometric transform\nmaps data\nsimplistic sketch\noriginal model in our minds\ndim image in a mirror\nunderstand\ndon’t\nnarrower\nmapping training inputs to training targets, point by point\ndeviates\nbreak in absurd ways\nfundamental differences\ngeometric morphing from input to output\ndeep-learning models\nhumans\nonly\nthemselves\nembodied experience\nnature of the underlying representations\nfar more\nimmediate stimuli\nimmediate responses\ncomplex, abstract models\nsituation\nourselves\nother people\nanticipate different possible futures\nlong-term planning\nmerge\nnever experienced before\nhypotheticals\nexpand\nbeyond\nexperience directly\nabstraction\nreasoning\ndefining characteristic of human cognition\nadapt\nnovel\nsituations\nno new data at all\nsharp contrast\ndeep nets\nstops making sense\nnew inputs\neven slightly\ntraining time\ndense sampling of the input space\nreliable mapping\npower of abstraction\nphysical models\nderive\none or a few trials\nhuman body\nsafely navigate a city\nnet\ndie\nthousands of times\nrelearn\nwithout having to die even once\nfar\npredict\nheaded\n?\nlong-term\nmore like programs\nbeyond\ncurrently\ncloser\nblend\nform\ngeneralization\ndon’t expect any singularitarian robot apocalypse to ensue\npure fantasy\nlong series of profound misunderstandings of both intelligence and tech",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning.
Highlight Ratio: 0.2 | The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach
even given vast amounts of human-annotated data. Say
you could assemble a dataset of
millions—of
descriptions of the features of a software product
as well as the corresponding source code
Even with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase
anything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw
Even learning a sorting algorithm with a deep neural network is tremendously difficult
This is because
deep-learning
is just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform
A deep-learning model can be interpreted as a kind of program; but
most programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves
or, even if one exists, it may not be learnable
the corresponding geometric transform may be far too complex, or there may not be appropriate data available
Scaling up current
techniques by stacking more layers and using more
data can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent
most of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold
One real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions
and knowledge on
things around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means
when we’re able to somewhat successfully train a model to generate captions
we’re led to believe that the model “understands” the contents
Then we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions
this is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model
it’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some
filter
you can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding
a gibbon gradient, we can get a neural network to classify the panda as a gibbon
This evidences both the brittleness of these models and the deep difference between their
mapping and our
perception
deep-learning models don’t have any understanding of their input
in a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds
like a dim image in a mirror
never fall into the trap of believing that neural networks understand the task they perform—they don’t
They were trained on a different, far narrower task
that of mapping training inputs to training targets, point by point. Show them anything that deviates from their training
and they will break in absurd ways
There are fundamental differences between the
geometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience
there’s a basic difference in the nature of the underlying representations.
Humans are capable of far more than mapping immediate stimuli to immediate responses
We maintain complex, abstract models of our
situation
ourselves
and
other people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge
known concepts to represent something we’ve never experienced before
This ability to handle hypotheticals, to expand our mental model space
beyond what we can experience directly—to perform abstraction and reasoning—is
the defining characteristic of human cognition
an ability to adapt to novel
situations using
no new data at all
This stands in sharp contrast with
deep nets
mapping
inputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time
Consider
the problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping
In contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials
if you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city
the net would have to die
thousands of times
Dropped into a new city, the net would have to relearn
what it knows
On the other hand, humans are able to learn safe behaviors without having to die even once
thanks to our power of abstract modeling of hypothetical situations
we’re
far from human-level AI
Given what we know of how deep nets work, their limitations, and the current state of
research
can we predict where things are headed
?
long-term
Models will be more like programs and will have capabilities that go
beyond the continuous geometric transformations of the input data we currently work with. These programs will
be
closer to
abstract mental models
models will blend algorithmic modules
This global library and associated model-growing system will be able to achieve some form of
generalization
But don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology | with deep learning
many applications are completely out of reach
even given vast
data
reasoning—like programming
the scientific method
long-term planning
is out of reach
because
deep-learning
is just
simple
geometric transformations mapping one vector space into another
assuming
a learnable continuous transform
most programs can’t be expressed as deep-learning models
either there exists no
network that solves
or
the
transform may be
too complex, or there may not be
data
more layers
can only superficially palliate some
issues. It won’t solve
more fundamental problems that
most
programs
can’t be expressed as
geometric morphing of a data manifold
humans
tend
to project intent
when
a model
generate captions
we
believe
the model “understands”
Then we’re surprised when
slight departure
causes
completely absurd captions
anything that deviates from
training
will break in absurd ways
Humans
maintain complex, abstract models
use these
to anticipate
futures and perform long-term planning
merge
concepts to represent something
never experienced
hypotheticals
expand our
model space
beyond
experience
using
no
data at all
in sharp contrast with
deep nets
to
navigate a city
the net would
die
thousands of times
humans
learn
without having to die
once
don’t expect any
robot apocalypse to ensue: that’s pure fantasy, coming from
profound misunderstandings of
intelligence and tech | 223 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: You can’t just scale up deep learning until you end up with AGI. It is a fundamentally different kind of problem than narrow machine learning.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "The space of applications that can be implemented with deep learning is nearly infinite. And yet, many applications are completely out of reach\neven given vast amounts of human-annotated data. Say\nyou could assemble a dataset of\nmillions—of\ndescriptions of the features of a software product\nas well as the corresponding source code\nEven with this data, you could not train a deep-learning model to read a product description and generate the appropriate codebase\nanything that requires reasoning—like programming or applying the scientific method—long-term planning, and algorithmic data manipulation is out of reach for deep-learning models, no matter how much data you throw\nEven learning a sorting algorithm with a deep neural network is tremendously difficult\nThis is because\ndeep-learning\nis just a chain of simple, continuous geometric transformations mapping one vector space into another. All it can do is map one data manifold X into another manifold Y, assuming the existence of a learnable continuous transform\nA deep-learning model can be interpreted as a kind of program; but\nmost programs can’t be expressed as deep-learning models—for most tasks, either there exists no corresponding deep-neural network that solves\nor, even if one exists, it may not be learnable\nthe corresponding geometric transform may be far too complex, or there may not be appropriate data available\nScaling up current\ntechniques by stacking more layers and using more\ndata can only superficially palliate some of these issues. It won’t solve the more fundamental problems that deep-learning models are limited in what they can represent\nmost of the programs you may wish to learn can’t be expressed as a continuous geometric morphing of a data manifold\nOne real risk with contemporary AI is misinterpreting what deep-learning models do and overestimating their abilities. A fundamental feature of humans is our theory of mind: our tendency to project intentions\nand knowledge on\nthings around us. Drawing a smiley face on a rock suddenly makes it “happy”—in our minds. Applied to deep learning, this means\nwhen we’re able to somewhat successfully train a model to generate captions\nwe’re led to believe that the model “understands” the contents\nThen we’re surprised when any slight departure from the sort of images present in the training data causes the model to generate completely absurd captions\nthis is highlighted by adversarial examples, which are samples fed to a deep-learning network that are designed to trick the model\nit’s possible to do gradient ascent in input space to generate inputs that maximize the activation of some\nfilter\nyou can slightly modify an image in order to maximize the class prediction for a given class. By taking a picture of a panda and adding\na gibbon gradient, we can get a neural network to classify the panda as a gibbon\nThis evidences both the brittleness of these models and the deep difference between their\nmapping and our\nperception\ndeep-learning models don’t have any understanding of their input\nin a human sense. Our own understanding of images, sounds, and language is grounded in our sensorimotor experience as humans. Machine-learning models have no access to such experiences and thus can’t understand their inputs in a human-relatable way. By annotating large numbers of training examples to feed into our models, we get them to learn a geometric transform that maps data to human concepts on a specific set of examples, but this mapping is a simplistic sketch of the original model in our minds\nlike a dim image in a mirror\nnever fall into the trap of believing that neural networks understand the task they perform—they don’t\nThey were trained on a different, far narrower task\nthat of mapping training inputs to training targets, point by point. Show them anything that deviates from their training\nand they will break in absurd ways\nThere are fundamental differences between the\ngeometric morphing from input to output that deep-learning models do, and the way humans think and learn. It isn’t only the fact that humans learn by themselves from embodied experience\nthere’s a basic difference in the nature of the underlying representations.\nHumans are capable of far more than mapping immediate stimuli to immediate responses\nWe maintain complex, abstract models of our\nsituation\nourselves\nand\nother people, and can use these models to anticipate different possible futures and perform long-term planning. We can merge\nknown concepts to represent something we’ve never experienced before\nThis ability to handle hypotheticals, to expand our mental model space\nbeyond what we can experience directly—to perform abstraction and reasoning—is\nthe defining characteristic of human cognition\nan ability to adapt to novel\nsituations using\nno new data at all\nThis stands in sharp contrast with\ndeep nets\nmapping\ninputs to outputs performed by a deep net quickly stops making sense if new inputs differ even slightly from what the net saw at training time\nConsider\nthe problem of learning the appropriate launch parameters to get a rocket to land on the moon. If you used a deep net for this task and trained it using supervised learning or reinforcement learning, you’d have to feed it thousands or even millions of launch trials: you’d need to expose it to a dense sampling of the input space, in order for it to learn a reliable mapping\nIn contrast, as humans we can use our power of abstraction to come up with physical models—rocket science—and derive an exact solution that will land the rocket on the moon in one or a few trials\nif you developed a deep net controlling a human body, and you wanted it to learn to safely navigate a city\nthe net would have to die\nthousands of times\nDropped into a new city, the net would have to relearn\nwhat it knows\nOn the other hand, humans are able to learn safe behaviors without having to die even once\nthanks to our power of abstract modeling of hypothetical situations\nwe’re\nfar from human-level AI\nGiven what we know of how deep nets work, their limitations, and the current state of\nresearch\ncan we predict where things are headed\n?\nlong-term\nModels will be more like programs and will have capabilities that go\nbeyond the continuous geometric transformations of the input data we currently work with. These programs will\nbe\ncloser to\nabstract mental models\nmodels will blend algorithmic modules\nThis global library and associated model-growing system will be able to achieve some form of\ngeneralization\nBut don’t expect any singularitarian robot apocalypse to ensue: that’s pure fantasy, coming from a long series of profound misunderstandings of both intelligence and technology",
"role": "user"
},
{
"content": "with deep learning\nmany applications are completely out of reach\neven given vast\ndata\nreasoning—like programming\nthe scientific method\nlong-term planning\nis out of reach\nbecause\ndeep-learning\nis just\nsimple\ngeometric transformations mapping one vector space into another\nassuming\na learnable continuous transform\nmost programs can’t be expressed as deep-learning models\neither there exists no\nnetwork that solves\nor\nthe\ntransform may be\ntoo complex, or there may not be\ndata\nmore layers\ncan only superficially palliate some\nissues. It won’t solve\nmore fundamental problems that\nmost\nprograms\ncan’t be expressed as\ngeometric morphing of a data manifold\nhumans\ntend\nto project intent\nwhen\na model\ngenerate captions\nwe\nbelieve\nthe model “understands”\nThen we’re surprised when\nslight departure\ncauses\ncompletely absurd captions\nanything that deviates from\ntraining\nwill break in absurd ways\nHumans\nmaintain complex, abstract models\nuse these\nto anticipate\nfutures and perform long-term planning\nmerge\nconcepts to represent something\nnever experienced\nhypotheticals\nexpand our\nmodel space\nbeyond\nexperience\nusing\nno\ndata at all\nin sharp contrast with\ndeep nets\nto\nnavigate a city\nthe net would\ndie\nthousands of times\nhumans\nlearn\nwithout having to die\nonce\ndon’t expect any\nrobot apocalypse to ensue: that’s pure fantasy, coming from\nprofound misunderstandings of\nintelligence and tech",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Even it can be scanned, it can’t be modeled. | The second step, translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts of the brain and to track and measure all the associated ion flux and other biochemical and signalling events, all of which is technically impossible. But even if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised. This is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells, each of which contributes to the experience by exhibiting a certain behaviour which depends on its molecular configuration.
For each cell, this configuration depends on (i) the genome it carries in its nucleus, (ii) the epigenetic chemical modifications of the genome (the DNA and its carrier molecules), and (iii) the molecular configuration (involving some hundreds of thousands of molecules) of the non-nuclear parts of the cell—all of which have been shaped by the experiences of the organism since conception.8
We cannot obtain a view of the interior structure of the cell (its molecular configuration) via imaging, because live imaging techniques do not have a sufficient resolution at either the spatial or the temporal level.9 * But even if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum; thus they measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not, however, improve our ability to emulate them.
In other words, we are not able to create a WBE; even if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond (which means we would obtain over 10^21 data points per second), we could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model.
And finally, there is the third step: simulation, which we do not discuss, since it relies for its input on the successful realisation of the first two steps, both of which are impossible.
The project of creating a superintelligent being by, in effect, copying without understanding, fails. But we note that, even if WBE were possible and a brainemulation without a body would indeed yield a ‘mind’, it would certainly not yield superintelligence, for it would at best emulate the intelligence of the copied mind. Bostrom explores how such ‘copies’ could then be ‘improved’, but we prefer to avert our readers’ eyes from the way he thereby adds yet more impossibilities to his pile. Instead, we have to face the fact that, when we die, the mind-body continuum ceases to exist, our mind ends, and six minutes after our last heartbeat or our last breath, our body has irreversibly lost the molecular configuration which enables it to function as an animate entity. It becomes a mere corpse—biological material that serves as nutrition for other organisms. | translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts
and to track and measure all the associated ion flux
even if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised
This is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells
For each cell, this configuration depends on
the genome
the epigenetic chemical modifications of the genome
and
the molecular configuration
all of which have been shaped by the experiences of the organism since conception
even if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum
they measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not
improve our ability to emulate them
we are not able to create a WBE
even if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond
which means we would obtain over 10^21 data points per second
we could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model
The project of creating a superintelligent being by
copying without understanding
fails | 224 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Even it can be scanned, it can’t be modeled.",
"role": "system"
},
{
"content": "The second step, translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts of the brain and to track and measure all the associated ion flux and other biochemical and signalling events, all of which is technically impossible. But even if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised. This is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells, each of which contributes to the experience by exhibiting a certain behaviour which depends on its molecular configuration.\nFor each cell, this configuration depends on (i) the genome it carries in its nucleus, (ii) the epigenetic chemical modifications of the genome (the DNA and its carrier molecules), and (iii) the molecular configuration (involving some hundreds of thousands of molecules) of the non-nuclear parts of the cell—all of which have been shaped by the experiences of the organism since conception.8\nWe cannot obtain a view of the interior structure of the cell (its molecular configuration) via imaging, because live imaging techniques do not have a sufficient resolution at either the spatial or the temporal level.9 * But even if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum; thus they measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not, however, improve our ability to emulate them.\nIn other words, we are not able to create a WBE; even if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond (which means we would obtain over 10^21 data points per second), we could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model.\nAnd finally, there is the third step: simulation, which we do not discuss, since it relies for its input on the successful realisation of the first two steps, both of which are impossible.\nThe project of creating a superintelligent being by, in effect, copying without understanding, fails. But we note that, even if WBE were possible and a brainemulation without a body would indeed yield a ‘mind’, it would certainly not yield superintelligence, for it would at best emulate the intelligence of the copied mind. Bostrom explores how such ‘copies’ could then be ‘improved’, but we prefer to avert our readers’ eyes from the way he thereby adds yet more impossibilities to his pile. Instead, we have to face the fact that, when we die, the mind-body continuum ceases to exist, our mind ends, and six minutes after our last heartbeat or our last breath, our body has irreversibly lost the molecular configuration which enables it to function as an animate entity. It becomes a mere corpse—biological material that serves as nutrition for other organisms.",
"role": "user"
},
{
"content": "translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts\nand to track and measure all the associated ion flux\neven if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised\nThis is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells\nFor each cell, this configuration depends on\nthe genome\nthe epigenetic chemical modifications of the genome\nand\nthe molecular configuration\nall of which have been shaped by the experiences of the organism since conception\neven if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum\nthey measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not\nimprove our ability to emulate them\nwe are not able to create a WBE\neven if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond\nwhich means we would obtain over 10^21 data points per second\nwe could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model\nThe project of creating a superintelligent being by\ncopying without understanding\nfails",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Even it can be scanned, it can’t be modeled. | translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts
and to track and measure all the associated ion flux
even if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised
This is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells
For each cell, this configuration depends on
the genome
the epigenetic chemical modifications of the genome
and
the molecular configuration
all of which have been shaped by the experiences of the organism since conception
even if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum
they measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not
improve our ability to emulate them
we are not able to create a WBE
even if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond
which means we would obtain over 10^21 data points per second
we could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model
The project of creating a superintelligent being by
copying without understanding
fails | translation
simultaneously measure
all neurotransmitters
all synaptic clefts
track
measure all
ion flux
even if we were to achieve all of this
still
emulate
relate them to the particular mental experiences of which consciousness is comprised
collaboration
many neurons
other (peripheral) cells
genome
epigenetic chemical modifications
molecular configuration
experiences of the organism since conception
do not have mathematical models to interpret the resulting data
evolve
complex system
mindbody continuum
directly
workings
inner experience
direct
would not
improve our ability to emulate them
create
even if we could observe
over 10^15 molecules
nanosecond
over 10^21 data points per second
could not emulate the behaviour of the collection of cells
no way to build a mathematical model that could do this
emulate
modeling it
complex systems
no way to produce the needed model
copying without understanding
fails | 225 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Even it can be scanned, it can’t be modeled.",
"role": "system"
},
{
"content": "translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts\nand to track and measure all the associated ion flux\neven if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised\nThis is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells\nFor each cell, this configuration depends on\nthe genome\nthe epigenetic chemical modifications of the genome\nand\nthe molecular configuration\nall of which have been shaped by the experiences of the organism since conception\neven if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum\nthey measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not\nimprove our ability to emulate them\nwe are not able to create a WBE\neven if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond\nwhich means we would obtain over 10^21 data points per second\nwe could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model\nThe project of creating a superintelligent being by\ncopying without understanding\nfails",
"role": "user"
},
{
"content": "translation\nsimultaneously measure\nall neurotransmitters\nall synaptic clefts\ntrack\nmeasure all\nion flux\neven if we were to achieve all of this\nstill\nemulate\nrelate them to the particular mental experiences of which consciousness is comprised\ncollaboration\nmany neurons\nother (peripheral) cells\ngenome\nepigenetic chemical modifications\nmolecular configuration\nexperiences of the organism since conception\ndo not have mathematical models to interpret the resulting data\nevolve\ncomplex system\nmindbody continuum\ndirectly\nworkings\ninner experience\ndirect\nwould not\nimprove our ability to emulate them\ncreate\neven if we could observe\nover 10^15 molecules\nnanosecond\nover 10^21 data points per second\ncould not emulate the behaviour of the collection of cells\nno way to build a mathematical model that could do this\nemulate\nmodeling it\ncomplex systems\nno way to produce the needed model\ncopying without understanding\nfails",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Even it can be scanned, it can’t be modeled.
Highlight Ratio: 0.3 | translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts
and to track and measure all the associated ion flux
even if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised
This is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells
For each cell, this configuration depends on
the genome
the epigenetic chemical modifications of the genome
and
the molecular configuration
all of which have been shaped by the experiences of the organism since conception
even if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum
they measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not
improve our ability to emulate them
we are not able to create a WBE
even if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond
which means we would obtain over 10^21 data points per second
we could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model
The project of creating a superintelligent being by
copying without understanding
fails | would require
ability to simultaneously measure
all neurotransmitters in all synaptic clefts
and
track and measure all the
ion flux
even if we
achieve
this, we still could not emulate these
because mental experiences are
collaboration of many
cells
depends on
genome
epigenetic
modifications
and
molecular
all
shaped by
experiences
even if we had
imaging
we
do not have
models to interpret the
data
even if we could observe
over 10^15 molecules at a resolution of a nanosecond
over 10^21 data points per second
we could not emulate the behaviour of
cells, because
no
mathematical model
could do this | 226 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Even it can be scanned, it can’t be modeled.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "translation would require the ability to simultaneously measure the amount of all neurotransmitters in all synaptic clefts\nand to track and measure all the associated ion flux\neven if we were to achieve all of this, we still could not emulate these activities in such a way as to relate them to the particular mental experiences of which consciousness is comprised\nThis is because mental experiences are the result of the collaboration of many neurons and other (peripheral) cells\nFor each cell, this configuration depends on\nthe genome\nthe epigenetic chemical modifications of the genome\nand\nthe molecular configuration\nall of which have been shaped by the experiences of the organism since conception\neven if we had machines capable of imaging at the required resolutions, we still do not have mathematical models to interpret the resulting data. These data are measurements of processes which evolve in the complex system that is the mindbody continuum\nthey measure directly the workings of this system, some of whose emanations we can perceive via inner experience. Even such direct measurement of the system’s workings would not\nimprove our ability to emulate them\nwe are not able to create a WBE\neven if we could observe the activities of billions of neurons with a totality of over 10^15 molecules at a resolution of a nanosecond\nwhich means we would obtain over 10^21 data points per second\nwe could not emulate the behaviour of the collection of cells, because there is no way to build a mathematical model that could do this. One cannot emulate a system without modeling it, and in the case of complex systems there is no way to produce the needed model\nThe project of creating a superintelligent being by\ncopying without understanding\nfails",
"role": "user"
},
{
"content": "would require\nability to simultaneously measure\nall neurotransmitters in all synaptic clefts\nand\ntrack and measure all the\nion flux\neven if we\nachieve\nthis, we still could not emulate these\nbecause mental experiences are\ncollaboration of many\ncells\ndepends on\ngenome\nepigenetic\nmodifications\nand\nmolecular\nall\nshaped by\nexperiences\neven if we had\nimaging\nwe\ndo not have\nmodels to interpret the\ndata\neven if we could observe\nover 10^15 molecules at a resolution of a nanosecond\nover 10^21 data points per second\nwe could not emulate the behaviour of\ncells, because\nno\nmathematical model\ncould do this",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless. | A former seafarer he notes that much of the technology being touted in the market today is not actually new. “The reality is when I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,” he explains.
While the potential capability has been there for many years for automation of vessels Hojgaard questions the business case and economics of replacing the crew. “Why would you replace the human onboard – because you can?” he asks.
Supposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost. That would equal roughly $350,000 or around $7m over 20 years.
“It’s like a drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology.”
In the case of say the diesel engine he believes this cannot be operated autonomously due to the high degree of oversight and maintenance required.
It is not that Hojgaard is a Luddite, far from it, but he believes a realistic view needs to be taken of how the industry can benefit from developments in AI.
An example of area where rapid development is taking place is in broadband communications at sea. “Communications is one of the area that will see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office – what difference does it make if you’re in Antwerp or on a ship,” he says. | much of the technology being touted in the market today is not actually new
when I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”
Why would you replace the human onboard – because you can?”
Supposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost
It’s
a drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology
the diesel engine
cannot be operated autonomously due to the high degree of oversight and maintenance required
a realistic view needs to be taken of how the industry can benefit from developments in AI.
An example of area where rapid development is taking place is in broadband communications at sea. “Communications
will see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office | 227 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless.",
"role": "system"
},
{
"content": "A former seafarer he notes that much of the technology being touted in the market today is not actually new. “The reality is when I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,” he explains.\nWhile the potential capability has been there for many years for automation of vessels Hojgaard questions the business case and economics of replacing the crew. “Why would you replace the human onboard – because you can?” he asks.\nSupposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost. That would equal roughly $350,000 or around $7m over 20 years.\n“It’s like a drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology.”\nIn the case of say the diesel engine he believes this cannot be operated autonomously due to the high degree of oversight and maintenance required.\nIt is not that Hojgaard is a Luddite, far from it, but he believes a realistic view needs to be taken of how the industry can benefit from developments in AI.\nAn example of area where rapid development is taking place is in broadband communications at sea. “Communications is one of the area that will see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office – what difference does it make if you’re in Antwerp or on a ship,” he says.",
"role": "user"
},
{
"content": "much of the technology being touted in the market today is not actually new\nwhen I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”\nWhy would you replace the human onboard – because you can?”\nSupposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost\nIt’s\na drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology\nthe diesel engine\ncannot be operated autonomously due to the high degree of oversight and maintenance required\na realistic view needs to be taken of how the industry can benefit from developments in AI.\nAn example of area where rapid development is taking place is in broadband communications at sea. “Communications\nwill see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless. | much of the technology being touted in the market today is not actually new
when I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”
Why would you replace the human onboard – because you can?”
Supposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost
It’s
a drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology
the diesel engine
cannot be operated autonomously due to the high degree of oversight and maintenance required
a realistic view needs to be taken of how the industry can benefit from developments in AI.
An example of area where rapid development is taking place is in broadband communications at sea. “Communications
will see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office | not actually new
conceivably
turns
there
20 years
Why would you replace the human onboard – because you can?”
current
21
eight
sophisticated
future
savings
one third of the crew cost
drop in the ocean
what it would cost to automate everything, design wise and changing some of the basic technology
diesel engine
cannot
oversight
maintenance
realistic view | 228 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless.",
"role": "system"
},
{
"content": "much of the technology being touted in the market today is not actually new\nwhen I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”\nWhy would you replace the human onboard – because you can?”\nSupposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost\nIt’s\na drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology\nthe diesel engine\ncannot be operated autonomously due to the high degree of oversight and maintenance required\na realistic view needs to be taken of how the industry can benefit from developments in AI.\nAn example of area where rapid development is taking place is in broadband communications at sea. “Communications\nwill see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office",
"role": "user"
},
{
"content": "not actually new\nconceivably\nturns\nthere\n20 years\nWhy would you replace the human onboard – because you can?”\ncurrent\n21\neight\nsophisticated\nfuture\nsavings\none third of the crew cost\ndrop in the ocean\nwhat it would cost to automate everything, design wise and changing some of the basic technology\ndiesel engine\ncannot\noversight\nmaintenance\nrealistic view",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless.
Highlight Ratio: 0.4 | much of the technology being touted in the market today is not actually new
when I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”
Why would you replace the human onboard – because you can?”
Supposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost
It’s
a drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology
the diesel engine
cannot be operated autonomously due to the high degree of oversight and maintenance required
a realistic view needs to be taken of how the industry can benefit from developments in AI.
An example of area where rapid development is taking place is in broadband communications at sea. “Communications
will see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office | much
tech
in the market
is not
new
20 years ago
the ship c
make all
turns itself
Why would you replace the human onboard – because you can?
Supposing the crew
can be reduced from
21 to eight
savings would
be
one third of the crew cost
It’s
a drop in the ocean compared to what it would cost to automate everything, design wise and changing
the basic technology
the diesel engine
cannot be operated autonomously due to
high
oversight and maintenance required
a realistic view needs to be taken | 229 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: It’s way cheaper to replace MOST of the crew and have them specialize. Full automation is an order of magnitude more expensive and pointless.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "much of the technology being touted in the market today is not actually new\nwhen I sailed 20 years ago and was a chief officer of containership with Maersk the ship could conceivably go from pilot station to pilot station by itself, it could make all the turns itself. The technology was there for 20 years for a lot of this stuff,”\nWhy would you replace the human onboard – because you can?”\nSupposing the crew compliment can be reduced from its current number of 21 to eight, with the higher level of education and training those eight would need to operate the sophisticated vessel of the future he argues savings would only be about one third of the crew cost\nIt’s\na drop in the ocean compared to what it would cost to automate everything, design wise and changing some of the basic technology\nthe diesel engine\ncannot be operated autonomously due to the high degree of oversight and maintenance required\na realistic view needs to be taken of how the industry can benefit from developments in AI.\nAn example of area where rapid development is taking place is in broadband communications at sea. “Communications\nwill see a huge leap forward in the next five years. In five years from now a ship will be as connected as any other office",
"role": "user"
},
{
"content": "much\ntech\nin the market\nis not\nnew\n20 years ago\nthe ship c\nmake all\nturns itself\nWhy would you replace the human onboard – because you can?\nSupposing the crew\ncan be reduced from\n21 to eight\nsavings would\nbe\none third of the crew cost\nIt’s\na drop in the ocean compared to what it would cost to automate everything, design wise and changing\nthe basic technology\nthe diesel engine\ncannot be operated autonomously due to\nhigh\noversight and maintenance required\na realistic view needs to be taken",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Experimentation is the best approach, which requires federalism | Several features of the existing occupational licensing system impede access to health care without providing appreciable protections for patients. Licensing restrictions prevent health care providers from offering services to the full extent of their competency, obstruct the adoption of telehealth, and deter foreign-trained providers from practicing in the United States. Scholars and policymakers have proposed a number of reforms to this system over the years, but these proposals have had a limited impact for political and institutional reasons.
Still, there are grounds for optimism. In recent years, the federal government has taken a range of initial steps to reform licensing requirements for health care providers, and these steps have the potential to improve access to health care. Together, they illustrate a federalist approach to licensing reform, in which the federal government encourages the states to reform their licensing regimes, while largely preserving states' control over the system. These steps include: (1) easing federal licensing restrictions for health care providers in certain areas where the federal government possesses regulatory authority; (2) creating incentives for states and professional bodies to experiment with reforms; (3) intensifying the Federal Trade Commission's focus on licensing boards' anti-competitive conduct; and (4) generating additional pressure for state-level reforms through expanding health insurance and promoting delivery system reforms under the Affordable Care Act.
This article argues that a federalist approach represents the most promising path toward reforming occupational licensing in health care. Federal intervention in licensing is necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of their licensing regimes, and their inability to resolve their own collective action problems. Nevertheless, large-scale federal preemption of state licensing laws is unlikely, due to a combination of interest group politics, Congress's tendency toward incrementalism, and its reliance on the states to administer federal policies. A federalist approach also has functional advantages over outright federal preemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations. Finally, this approach presents a model for how the federal government can play a constructive role in occupational licensing in other fields besides health care, and in other areas of state regulatory policy. | a federalist approach to licensing reform, in which the federal government encourages
states to reform
licensing
while largely preserving states' control
represents the most promising path toward reforming occupational licensing
Federal intervention
is necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of
licensing
and their inability to resolve their own collective action problems
A federalist approach
has functional advantages over outright
preemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations
this
presents a model for how the federal government can play a constructive role in
other fields
of state regulatory policy. | 230 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Experimentation is the best approach, which requires federalism",
"role": "system"
},
{
"content": "Several features of the existing occupational licensing system impede access to health care without providing appreciable protections for patients. Licensing restrictions prevent health care providers from offering services to the full extent of their competency, obstruct the adoption of telehealth, and deter foreign-trained providers from practicing in the United States. Scholars and policymakers have proposed a number of reforms to this system over the years, but these proposals have had a limited impact for political and institutional reasons.\nStill, there are grounds for optimism. In recent years, the federal government has taken a range of initial steps to reform licensing requirements for health care providers, and these steps have the potential to improve access to health care. Together, they illustrate a federalist approach to licensing reform, in which the federal government encourages the states to reform their licensing regimes, while largely preserving states' control over the system. These steps include: (1) easing federal licensing restrictions for health care providers in certain areas where the federal government possesses regulatory authority; (2) creating incentives for states and professional bodies to experiment with reforms; (3) intensifying the Federal Trade Commission's focus on licensing boards' anti-competitive conduct; and (4) generating additional pressure for state-level reforms through expanding health insurance and promoting delivery system reforms under the Affordable Care Act.\nThis article argues that a federalist approach represents the most promising path toward reforming occupational licensing in health care. Federal intervention in licensing is necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of their licensing regimes, and their inability to resolve their own collective action problems. Nevertheless, large-scale federal preemption of state licensing laws is unlikely, due to a combination of interest group politics, Congress's tendency toward incrementalism, and its reliance on the states to administer federal policies. A federalist approach also has functional advantages over outright federal preemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations. Finally, this approach presents a model for how the federal government can play a constructive role in occupational licensing in other fields besides health care, and in other areas of state regulatory policy.",
"role": "user"
},
{
"content": "a federalist approach to licensing reform, in which the federal government encourages\nstates to reform\nlicensing\nwhile largely preserving states' control\nrepresents the most promising path toward reforming occupational licensing\nFederal intervention\nis necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of\nlicensing\nand their inability to resolve their own collective action problems\nA federalist approach\nhas functional advantages over outright\npreemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations\nthis\npresents a model for how the federal government can play a constructive role in\nother fields\nof state regulatory policy.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Experimentation is the best approach, which requires federalism | a federalist approach to licensing reform, in which the federal government encourages
states to reform
licensing
while largely preserving states' control
represents the most promising path toward reforming occupational licensing
Federal intervention
is necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of
licensing
and their inability to resolve their own collective action problems
A federalist approach
has functional advantages over outright
preemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations
this
presents a model for how the federal government can play a constructive role in
other fields
of state regulatory policy. | federalist approach
encourages
states
reform
licensing
preserving states' control
most promising path
Federal intervention
is necessary
incentives to experiment
externalities
collective action problems
federalist
functional advantages
outright
preemption
more experimentation in constructing new licensing regimes
f
g
advantage of states' institutional expertise in regulating occupations
model
f
g
constructive role in
other fields
of state regulatory policy. | 231 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Experimentation is the best approach, which requires federalism",
"role": "system"
},
{
"content": "a federalist approach to licensing reform, in which the federal government encourages\nstates to reform\nlicensing\nwhile largely preserving states' control\nrepresents the most promising path toward reforming occupational licensing\nFederal intervention\nis necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of\nlicensing\nand their inability to resolve their own collective action problems\nA federalist approach\nhas functional advantages over outright\npreemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations\nthis\npresents a model for how the federal government can play a constructive role in\nother fields\nof state regulatory policy.",
"role": "user"
},
{
"content": "federalist approach\nencourages\nstates\nreform\nlicensing\npreserving states' control\nmost promising path\nFederal intervention\nis necessary\nincentives to experiment\nexternalities\ncollective action problems\nfederalist\nfunctional advantages\noutright\npreemption\nmore experimentation in constructing new licensing regimes\nf\ng\nadvantage of states' institutional expertise in regulating occupations\nmodel\nf\ng\nconstructive role in\nother fields\nof state regulatory policy.",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Experimentation is the best approach, which requires federalism
Highlight Ratio: 0.5 | a federalist approach to licensing reform, in which the federal government encourages
states to reform
licensing
while largely preserving states' control
represents the most promising path toward reforming occupational licensing
Federal intervention
is necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of
licensing
and their inability to resolve their own collective action problems
A federalist approach
has functional advantages over outright
preemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations
this
presents a model for how the federal government can play a constructive role in
other fields
of state regulatory policy. | a federalist approach
encourages
states to reform
while
preserving states' control
A federalist approach
has functional advantages over outright
preemption: it allows
more experimentation in constructing new licensing regimes, and
enables the f
g
to take advantage of states' institutional expertise in regulating occupations
presents a model for how
f
g
can play a constructive role in
state regulatory policy | 232 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Experimentation is the best approach, which requires federalism\n\nHighlight Ratio: 0.5",
"role": "system"
},
{
"content": "a federalist approach to licensing reform, in which the federal government encourages\nstates to reform\nlicensing\nwhile largely preserving states' control\nrepresents the most promising path toward reforming occupational licensing\nFederal intervention\nis necessary, due to states' lack of incentives to experiment with licensing reforms, the externalities of\nlicensing\nand their inability to resolve their own collective action problems\nA federalist approach\nhas functional advantages over outright\npreemption: it allows for more experimentation in constructing new licensing regimes, and it enables the federal government to take advantage of states' institutional expertise in regulating occupations\nthis\npresents a model for how the federal government can play a constructive role in\nother fields\nof state regulatory policy.",
"role": "user"
},
{
"content": "a federalist approach\nencourages\nstates to reform\nwhile\npreserving states' control\nA federalist approach\nhas functional advantages over outright\npreemption: it allows\nmore experimentation in constructing new licensing regimes, and\nenables the f\ng\nto take advantage of states' institutional expertise in regulating occupations\npresents a model for how\nf\ng\ncan play a constructive role in\nstate regulatory policy",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Kills regulatory innovation | 3. Checks and Balances
Finally, and perhaps most importantly, the Framers of the Constitution envisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled. 226 Permitting the FTC or private plaintiffs to unduly interfere with a state's core police powers would denigrate the federalist structure guaranteed by the Constitution and affirmed by the Tenth Amendment.227 Indeed, a state's right to regulate its internal economic affairs is at the core of its police powers, and permitting federal government interference with the exercise of these powers "would turn federalism on its head." 228 Although the problems caused by excessive occupational licensing requirements offend fundamental notions of economic liberty and free markets, antitrust federalism should not be abandoned wholesale when it occasionally, or even frequently, facilitates bad results. | the Framers
envisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled
Permitting the
to unduly interfere with a state's core police powers would denigrate the federalist structure
a state's right to regulate its internal economic affairs is at the core of its police powers
and permitting federal government interference with the exercise
would turn federalism on its head
antitrust federalism should not be abandoned wholesale | 233 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Kills regulatory innovation",
"role": "system"
},
{
"content": "3. Checks and Balances\nFinally, and perhaps most importantly, the Framers of the Constitution envisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled. 226 Permitting the FTC or private plaintiffs to unduly interfere with a state's core police powers would denigrate the federalist structure guaranteed by the Constitution and affirmed by the Tenth Amendment.227 Indeed, a state's right to regulate its internal economic affairs is at the core of its police powers, and permitting federal government interference with the exercise of these powers \"would turn federalism on its head.\" 228 Although the problems caused by excessive occupational licensing requirements offend fundamental notions of economic liberty and free markets, antitrust federalism should not be abandoned wholesale when it occasionally, or even frequently, facilitates bad results.",
"role": "user"
},
{
"content": "the Framers\nenvisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled\nPermitting the\nto unduly interfere with a state's core police powers would denigrate the federalist structure\na state's right to regulate its internal economic affairs is at the core of its police powers\nand permitting federal government interference with the exercise\nwould turn federalism on its head\nantitrust federalism should not be abandoned wholesale",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Kills regulatory innovation | the Framers
envisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled
Permitting the
to unduly interfere with a state's core police powers would denigrate the federalist structure
a state's right to regulate its internal economic affairs is at the core of its police powers
and permitting federal government interference with the exercise
would turn federalism on its head
antitrust federalism should not be abandoned wholesale | states retain the authority to which they are entitled
unduly interfere with a state's core police powers
denigrate the federalist structure
internal
core of its police powers
interference
turn federalism on its head
not be abandoned wholesale | 234 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Kills regulatory innovation",
"role": "system"
},
{
"content": "the Framers\nenvisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled\nPermitting the\nto unduly interfere with a state's core police powers would denigrate the federalist structure\na state's right to regulate its internal economic affairs is at the core of its police powers\nand permitting federal government interference with the exercise\nwould turn federalism on its head\nantitrust federalism should not be abandoned wholesale",
"role": "user"
},
{
"content": "states retain the authority to which they are entitled\nunduly interfere with a state's core police powers\ndenigrate the federalist structure\ninternal\ncore of its police powers\ninterference\nturn federalism on its head\nnot be abandoned wholesale",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Kills regulatory innovation
Highlight Ratio: 0.7 | the Framers
envisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled
Permitting the
to unduly interfere with a state's core police powers would denigrate the federalist structure
a state's right to regulate its internal economic affairs is at the core of its police powers
and permitting federal government interference with the exercise
would turn federalism on its head
antitrust federalism should not be abandoned wholesale | Framers
envisioned
checks and balances
central to
this
is that
states retain the authority to which they are entitled
Permitting
to
interfere with a state's
police powers would denigrate the federalist structure
to regulate
internal economic affairs is
the core of
police powers
interference
would turn federalism on its head
antitrust federalism should not be abandoned | 235 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Kills regulatory innovation\n\nHighlight Ratio: 0.7",
"role": "system"
},
{
"content": "the Framers\nenvisioned a system of checks and balances; central to a government operating under this ideal is that the states retain the authority to which they are entitled\nPermitting the\nto unduly interfere with a state's core police powers would denigrate the federalist structure\na state's right to regulate its internal economic affairs is at the core of its police powers\nand permitting federal government interference with the exercise\nwould turn federalism on its head\nantitrust federalism should not be abandoned wholesale",
"role": "user"
},
{
"content": "Framers\nenvisioned\nchecks and balances\ncentral to\nthis\nis that\nstates retain the authority to which they are entitled\nPermitting\nto\ninterfere with a state's\npolice powers would denigrate the federalist structure\nto regulate\ninternal economic affairs is\nthe core of\npolice powers\ninterference\nwould turn federalism on its head\nantitrust federalism should not be abandoned",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Preserving some state authority is vital for state innovation and flexibility. | Preserving state variation in licensing policy would also address the concern that Congress doesn't have enough information to enact the optimal licensing reform. 304 Although proponents of federal preemption may object that such an approach is slower and less efficient than having Congress step in and set federal standards, 305 the potential downside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards. 306 For instance, if the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements. 307
Finally, a federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level. Daniel Gilman, an Attorney Advisor at the FTC's Office of Policy Planning, notes that there is currently no federal agency "with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one." 308 He acknowledges that this challenge is not insurmountable: there is plenty of expertise about occupational regulation in the federal government, at agencies like the Department of Labor, HRSA, and the FTC. But some federal preemption proposals, such as creating a federal system of licensure, would necessitate creating [*353] an accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws. 309
* * *
All this is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it. 310 Simply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes. 311 In addition, as outlined above, a "field-claiming" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.
V. CONCLUSION
This article presents a model for how the federal government can encourage states to reform the current system of occupational licensing for health care providers, and argues that such an approach would be more viable than either leaving licensing to the states or advocating for outright federal preemption. Although this paper focuses primarily on occupational licensing in health care, it also suggests that a federalist model could be used to change the licensing systems governing other fields as well. 312
[*354] Of course, that is not to say that affecting widespread change in our licensing system will be easy or quick. The licensing system in the United States has proven remarkably intractable since its inception, and any change will likely have to be incremental. Nearly half a century ago, in their study of occupational licensing, Benjamin Shimberg, Barbara F. Esser, and Daniel H. Kruger sketched out some of the challenges entailed in reforming the licensing system:
[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups that have, over the years, managed to achieve some sort of delicate balance within the existing structure. The possibility of change, even relatively minor change, is likely to be perceived as a threat by those who gain not only prestige but also tangible economic benefits from the existing structure. Anyone contemplating change must consider not only its operational aspects, such as amending existing legislation or modifying procedures, but also its psychological aspects--the way people perceive or respond to the proposed changes. It is probably best to think of modifications in licensing as an ongoing process--a spiral moving upward from one level to the next--that will not necessarily be accomplished in one, two, or even five years. 313
Despite the promise of recent developments and the resurgence of interest in licensing, this cautionary advice seems just as applicable today. Any large-scale change will require a strong and sustained commitment on behalf of the federal government, and will likely take place over a long period of time.
Nevertheless, this article provides some reasons for optimism. It shows that the federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements for health care providers specifically, and that even these limited measures have had tangible impacts on states' licensing regimes and in turn have the potential to improve access to health care.
This lesson is especially important today, as there is a growing appreciation of the important consequences that state and local regulations have for national issues such as health care, the labor market, and geographic mobility. Many scholars and policymakers grappling with these interactions have continued to view state and local regulatory policy through a dual federalist framework: either advocating [*355] for states to reform their own regulations, or alternatively, calling for large-scale federal preemption. By contrast, this article shows that at least in some contexts, a federalist solution may be the most realistic and normatively desirable solution, and that there is promising precedent for such an approach. | Congress doesn't have enough information to enact the optimal licensing reform
Although proponents of federal preemption may object that such an approach is slower and less efficient
the
downside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards
if the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements
a federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level
there is currently no federal agency "with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one."
this challenge is not insurmountable
But
federal preemption
such as creating a federal system of licensure, would necessitate creating
an accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws
this is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it
Simply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes
a "field-claiming" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.
[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups
Nevertheless
the federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements
even these limited measures have had tangible impacts
and in turn have the potential to improve access
a federalist solution may be the most realistic and normatively desirable solution
there is promising precedent for such an approach. | 236 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Preserving some state authority is vital for state innovation and flexibility.",
"role": "system"
},
{
"content": "Preserving state variation in licensing policy would also address the concern that Congress doesn't have enough information to enact the optimal licensing reform. 304 Although proponents of federal preemption may object that such an approach is slower and less efficient than having Congress step in and set federal standards, 305 the potential downside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards. 306 For instance, if the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements. 307\nFinally, a federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level. Daniel Gilman, an Attorney Advisor at the FTC's Office of Policy Planning, notes that there is currently no federal agency \"with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one.\" 308 He acknowledges that this challenge is not insurmountable: there is plenty of expertise about occupational regulation in the federal government, at agencies like the Department of Labor, HRSA, and the FTC. But some federal preemption proposals, such as creating a federal system of licensure, would necessitate creating [*353] an accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws. 309\n* * *\nAll this is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it. 310 Simply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes. 311 In addition, as outlined above, a \"field-claiming\" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.\nV. CONCLUSION\nThis article presents a model for how the federal government can encourage states to reform the current system of occupational licensing for health care providers, and argues that such an approach would be more viable than either leaving licensing to the states or advocating for outright federal preemption. Although this paper focuses primarily on occupational licensing in health care, it also suggests that a federalist model could be used to change the licensing systems governing other fields as well. 312\n [*354] Of course, that is not to say that affecting widespread change in our licensing system will be easy or quick. The licensing system in the United States has proven remarkably intractable since its inception, and any change will likely have to be incremental. Nearly half a century ago, in their study of occupational licensing, Benjamin Shimberg, Barbara F. Esser, and Daniel H. Kruger sketched out some of the challenges entailed in reforming the licensing system:\n[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups that have, over the years, managed to achieve some sort of delicate balance within the existing structure. The possibility of change, even relatively minor change, is likely to be perceived as a threat by those who gain not only prestige but also tangible economic benefits from the existing structure. Anyone contemplating change must consider not only its operational aspects, such as amending existing legislation or modifying procedures, but also its psychological aspects--the way people perceive or respond to the proposed changes. It is probably best to think of modifications in licensing as an ongoing process--a spiral moving upward from one level to the next--that will not necessarily be accomplished in one, two, or even five years. 313\nDespite the promise of recent developments and the resurgence of interest in licensing, this cautionary advice seems just as applicable today. Any large-scale change will require a strong and sustained commitment on behalf of the federal government, and will likely take place over a long period of time.\nNevertheless, this article provides some reasons for optimism. It shows that the federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements for health care providers specifically, and that even these limited measures have had tangible impacts on states' licensing regimes and in turn have the potential to improve access to health care.\nThis lesson is especially important today, as there is a growing appreciation of the important consequences that state and local regulations have for national issues such as health care, the labor market, and geographic mobility. Many scholars and policymakers grappling with these interactions have continued to view state and local regulatory policy through a dual federalist framework: either advocating [*355] for states to reform their own regulations, or alternatively, calling for large-scale federal preemption. By contrast, this article shows that at least in some contexts, a federalist solution may be the most realistic and normatively desirable solution, and that there is promising precedent for such an approach.",
"role": "user"
},
{
"content": "Congress doesn't have enough information to enact the optimal licensing reform\nAlthough proponents of federal preemption may object that such an approach is slower and less efficient\nthe\ndownside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards\nif the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements\na federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level\nthere is currently no federal agency \"with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one.\"\nthis challenge is not insurmountable\nBut\nfederal preemption\nsuch as creating a federal system of licensure, would necessitate creating\nan accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws\nthis is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it\nSimply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes\na \"field-claiming\" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.\n[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups\nNevertheless\nthe federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements\neven these limited measures have had tangible impacts\nand in turn have the potential to improve access\na federalist solution may be the most realistic and normatively desirable solution\nthere is promising precedent for such an approach.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Preserving some state authority is vital for state innovation and flexibility. | Congress doesn't have enough information to enact the optimal licensing reform
Although proponents of federal preemption may object that such an approach is slower and less efficient
the
downside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards
if the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements
a federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level
there is currently no federal agency "with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one."
this challenge is not insurmountable
But
federal preemption
such as creating a federal system of licensure, would necessitate creating
an accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws
this is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it
Simply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes
a "field-claiming" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.
[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups
Nevertheless
the federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements
even these limited measures have had tangible impacts
and in turn have the potential to improve access
a federalist solution may be the most realistic and normatively desirable solution
there is promising precedent for such an approach. | Congress doesn't have enough information to enact the optimal licensing reform
slower
less efficient
downside
federal
nationwide
also bigger
no guarantee
f
g
optimal standards
could
reduce access
strictest
federalist
institutional experience
equivalent institutional apparatus at the federal level
no federal agency
expertise
difficult to create one
insurmountable
not
no
incompatible
coop
threat
option
or
federal government
generalized threat
impetus
states
better served
several tools
prod
their
hundreds of state laws
local ordinances
dozens of occupational interest groups
range of tools at its disposal
already
limited
tangible impacts
federalist
realistic
normatively desirable
promising precedent | 237 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Preserving some state authority is vital for state innovation and flexibility.",
"role": "system"
},
{
"content": "Congress doesn't have enough information to enact the optimal licensing reform\nAlthough proponents of federal preemption may object that such an approach is slower and less efficient\nthe\ndownside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards\nif the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements\na federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level\nthere is currently no federal agency \"with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one.\"\nthis challenge is not insurmountable\nBut\nfederal preemption\nsuch as creating a federal system of licensure, would necessitate creating\nan accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws\nthis is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it\nSimply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes\na \"field-claiming\" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.\n[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups\nNevertheless\nthe federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements\neven these limited measures have had tangible impacts\nand in turn have the potential to improve access\na federalist solution may be the most realistic and normatively desirable solution\nthere is promising precedent for such an approach.",
"role": "user"
},
{
"content": "Congress doesn't have enough information to enact the optimal licensing reform\nslower\nless efficient\ndownside\nfederal\nnationwide\nalso bigger\nno guarantee\nf\ng\noptimal standards\ncould\nreduce access\nstrictest\nfederalist\ninstitutional experience\nequivalent institutional apparatus at the federal level\nno federal agency\nexpertise\ndifficult to create one\ninsurmountable\nnot\nno\nincompatible\ncoop\nthreat\noption\nor\nfederal government\ngeneralized threat\nimpetus\nstates\nbetter served\nseveral tools\nprod\ntheir\nhundreds of state laws\nlocal ordinances\ndozens of occupational interest groups\nrange of tools at its disposal\nalready\nlimited\ntangible impacts\nfederalist\nrealistic\nnormatively desirable\npromising precedent",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Preserving some state authority is vital for state innovation and flexibility.
Highlight Ratio: 0.2 | Congress doesn't have enough information to enact the optimal licensing reform
Although proponents of federal preemption may object that such an approach is slower and less efficient
the
downside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards
if the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements
a federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level
there is currently no federal agency "with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one."
this challenge is not insurmountable
But
federal preemption
such as creating a federal system of licensure, would necessitate creating
an accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws
this is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it
Simply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes
a "field-claiming" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.
[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups
Nevertheless
the federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements
even these limited measures have had tangible impacts
and in turn have the potential to improve access
a federalist solution may be the most realistic and normatively desirable solution
there is promising precedent for such an approach. | Congress doesn't have
information to enact
optimal licensing
the
downside of
federal
standards
nationwide is
bigger, since there is no guarantee
the f
g
will set
optimal standards
federal
standards
could actually reduce access
a federalist approach
enable
to take advantage of states' institutional experience in regulating
and
avoid
an equivalent
apparatus at the federal level
there is
no federal agency "with the
expertise
it would be difficult to create one
not
that there is no role
In fact
coop
regimes
made use of the threat of
preemption
the generalized threat
provide
impetus for states to reform
a federalist solution may be
most
normatively desirable | 238 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Preserving some state authority is vital for state innovation and flexibility.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "Congress doesn't have enough information to enact the optimal licensing reform\nAlthough proponents of federal preemption may object that such an approach is slower and less efficient\nthe\ndownside of having the federal government pick certain standards and apply them nationwide is also bigger, since there is no guarantee that the federal government will set the optimal standards\nif the federal government were to adopt federal scope-of practice standards, it could actually reduce access to care to the extent that the federal scope-of-practice standards were set to match the strictest state requirements\na federalist approach would enable the federal government to take advantage of states' institutional experience in regulating occupations, and to avoid having to create an equivalent institutional apparatus at the federal level\nthere is currently no federal agency \"with the authority, expertise, and experience to perform the various licensing functions undertaken by the states, and it would be difficult to create one.\"\nthis challenge is not insurmountable\nBut\nfederal preemption\nsuch as creating a federal system of licensure, would necessitate creating\nan accompanying set of institutions capable of issuing, updating, and enforcing federal occupational licensing laws\nthis is not to say that there is no role for federal preemption, nor that federal preemption is necessarily incompatible with a federalist approach. In fact, traditional cooperative federalist regimes have often made use of the threat of federal preemption: offering states the option of administering a certain program or having the federal government administer it\nSimply maintaining the generalized threat of federal preemption may provide additional impetus for states to reform their licensing regimes\na \"field-claiming\" approach, in which the federal government makes a small series of encroachments onto state territory, may represent a realistic path. Yet reformers would be better served by viewing preemption as one of several tools that can be used to prod the states into reforming their licensing regimes, rather than as an immediate or exclusive solution.\n[T]he whole institution of occupational licensing is embedded in a morass of federal, state, and local legislation suffused with tradition, custom, and jealously guarded rights. There are clearly no simple solutions. To bring about change would involve not only modifications of hundreds of state laws and local ordinances but also negotiations among dozens of occupational interest groups\nNevertheless\nthe federal government has a range of tools at its disposal to encourage states to reform their licensing regimes, that it has already taken some initial steps to encourage states to restructure their licensing requirements\neven these limited measures have had tangible impacts\nand in turn have the potential to improve access\na federalist solution may be the most realistic and normatively desirable solution\nthere is promising precedent for such an approach.",
"role": "user"
},
{
"content": "Congress doesn't have\ninformation to enact\noptimal licensing\nthe\ndownside of\nfederal\nstandards\nnationwide is\nbigger, since there is no guarantee\nthe f\ng\nwill set\noptimal standards\nfederal\nstandards\ncould actually reduce access\na federalist approach\nenable\nto take advantage of states' institutional experience in regulating\nand\navoid\nan equivalent\napparatus at the federal level\nthere is\nno federal agency \"with the\nexpertise\nit would be difficult to create one\nnot\nthat there is no role\nIn fact\ncoop\nregimes\nmade use of the threat of\npreemption\nthe generalized threat\nprovide\nimpetus for states to reform\na federalist solution may be\nmost\nnormatively desirable",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.” | Inspector General Michael Horowitz’s report last week told a complex story about extraordinary events related to the investigation of officials in Donald Trump’s 2016 presidential campaign. Its publication predictably ignited a contest between Democrats and Republicans to extract from the 429-page opus what was most serviceable in the construction of competing political “narratives.”
But there is something much more important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed in how the FBI and the Department of Justice, of which it is a component, open and conduct investigations—especially those related to politicians and political campaigns. The report prompted concerns from both sides of the aisle, suggesting that there’s an opportunity for serious reflection and reform—if Congress and the executive branch can seize it.
The mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party over another, or to serve the FBI’s institutional interests at the expense of elected officials.
These tactics, when publicly disclosed in the 1960s and ’70s, almost destroyed the bureau. The FBI survived with its powerful investigatory and surveillance powers intact only because it entered into a “grand bargain” with Congress and the American people about the use of such powers.
The central elements of that bargain, as one of us previously explained, were as follows:
The president and his intelligence bureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions on how they collected, analyzed, and disseminated intelligence information; a bevy of lawyers throughout the intelligence community and, over time, in the Justice Department monitored and enforced those restrictions; domestic surveillance required a court order, including a court order from a new court, the Foreign Intelligence Surveillance Court, for foreign intelligence investigations; and two new committees, the Senate and House Intelligence committees, were to be kept “fully and currently informed” of all significant intelligence activities, and would have robust oversight authorities. The idea was that the use of these powers would be documented and watched by institutions that could be trusted to keep secrets but would act as credible surrogates for public oversight mechanisms.
The point of these arrangements was not only to assure the American people that the FBI (and other intelligence agencies) did not in fact use its powers for political ends, but also to ensure that it did not appear to use its powers for these ends.
The FBI has taken a large hit in its credibility over the past four years, due in large part to Trump’s unprecedented, reckless, and routinely baseless attacks on it. But the bureau has also hurt itself through its conduct of the investigation of Trump-campaign officials and of Hillary Clinton’s emails when she was a presidential candidate. Horowitz, in his latest report, and in three earlier ones related to the 2016 election, has documented what appear to be serious failures of policy, process, and decision making at multiple levels.
While Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee. Among Democrats’ numerous other concerns about the conduct of the Clinton email investigation, they were appalled by reasons to believe that a “a deep and visceral hatred” of Hillary Clinton in the FBI’s New York field office led agents to leak investigative information to Rudy Giuliani and Devin Nunes, and that the threat of further leaks informed the FBI’s decision to notify Congress, only 11 days before the 2016 election, that it had reopened the Clinton email investigation.
The problems at the FBI revealed by multiple inspector-general investigations do not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive or manipulation. The overall effect of these controversies has been to call into question the viability of the grand bargain, at least on its current terms, especially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But based on his insights and other factors, we think that three areas warrant special consideration.
First is the serious problem of inadequate guidance to the FBI for opening and conducting investigations against politicians or campaigns, especially during election season. The FBI has two guidebooks—the Attorney General’s Guidelines, and the Domestic Investigations and Operations Guide (DIOG)—to regulate the bureau’s investigative and intelligence-collection activities. Such guidance is vital to FBI legitimacy, especially when it investigates a politician or campaign. Well-designed guidance can establish a process to ensure that controversial steps are properly vetted in ways that minimize mistakes or disproportionate action. This, in turn, can help the FBI avoid controversy later, either because it declined to take unnecessarily controversial investigatory steps, or because it took those steps but can point to clear guidance for legitimation. But the past three years have shown that the current guidance is not well designed for investigations related to political campaigns.
As Horowitz noted, the threshold for opening an FBI investigation is “low.” To open a “full investigation” like the one against Trump-campaign officials—which allows the FBI to use all its prodigious surveillance and intelligence-gathering techniques—the FBI needs to satisfy only two requirements. First, it must have an “authorized purpose” (such as detecting a federal crime or threat to national security). And second, it must have an “articulable factual basis” that “reasonably indicates” that a federal crime or threat to national security “may have occurred” or “may occur,” and that the investigation may obtain relevant information. These relatively permissive criteria give officials enormous leeway in opening an investigation based merely on the right type of factual allegations. And there is only one additional rule when the investigation involves a political candidate or member of a political organization: The FBI agent considering such a step must consult up the FBI chain of command to various degrees depending on the type of assessment or investigation in issue.
Horowitz’s latest report expressed surprise about deficiencies in the DIOG guidance in the context of opening investigations related to political campaigns. He proposed that the FBI establish procedures that require giving notice not just to a senior FBI official, but also outside the bureau to a senior DOJ official, such as the deputy attorney general, “for case openings that implicate core First Amendment activity.” Such elevated notice at the outset of such a sensitive investigation is a good idea because it brings into the loop a Justice Department official who has been subject to Senate confirmation and is directly subject to congressional oversight, and thus politically accountable. Such high-level notice occurs in other crucial circumstances, such as sensitive undercover operations. A senior Senate-confirmed Justice Department official will also likely take a broader view of the matter than the Bureau and can, as Horowitz noted in his hearing, “consider the potential constitutional and prudential implications in advance of these activities.” In the event that such an official fails to do so, or makes the attempt but exercises poor judgment, he or she is the one who should answer for this to Congress and public.
We would go further and propose not just better notice rules up the chain of command, but also concrete guidance to FBI officials about when they can or cannot open cases in politically sensitive areas. Horowitz noted that the FBI rulebook uses the same low threshold for opening sensitive cases connected to a political campaign, which implicate First Amendment activity and electoral integrity, as in ordinary, nonsensitive cases. The FBI should develop a heightened threshold for opening such cases to ensure that First Amendment activity is not unduly investigated. It is a fair concern that the new standard could invite risk aversion within the bureau and leave undetected criminal or adverse national-security activity related to a campaign. This requires particularly close attention in an era likely to be marked by increased activity to illicitly influence our political process. But given the very large costs of the investigations in the past three years, these risks seem acceptable and can be mitigated if the heightened standard is carefully crafted.
We would also propose a heightened standard, and mandatory notice to the Justice Department, before the FBI opens a counterintelligence investigation on the president, as it did in May 2017. Such investigations focus on threats to national security from foreign intelligence agencies. The FBI is authorized to open such investigations but lacks specific guidance when the possible foreign asset is the president. The FBI appears to have concluded that Trump posed a national-security threat, and warranted counterintelligence scrutiny, because he was harming the FBI’s Russia investigation. But the president is not like other individuals who might pose a threat to national security, because he is constitutionally empowered to determine the national-security interests of the United States and to conduct its foreign policy, which includes the authority to change foreign policy on a dime.
We are not saying that the FBI should be barred from investigating a president who appears to be a Manchurian candidate, but the situation requires special guidance. The lines between a Manchurian candidate, a president engaged in quid pro quo foreign-policy bargaining that is possibly corrupt, and the very large discretionary control that presidents have to conduct U.S. foreign policy are not always easy to discern. The decision to open a counterintelligence investigation of the president is so consequential for the nation (and the FBI), and so politically fraught, that it should not be made without much clearer guidance. It also should not be made by the FBI alone. FBI leadership wisely realized this truth, even without specific guidance, because they quickly briefed the acting attorney general and relevant congressional leaders. Such briefings should be mandatory—both for the FBI’s sake, and to ensure that Justice Department leaders do not eschew their responsibilities.
Horowitz additionally noted that the FBI had no policy on whether to give notice to campaigns, like Trump’s, that are subject to investigation. According to Horowitz, the discretionary call not to inform the Trump campaign was made by an assistant director of an FBI division, who concluded without any guidance that such notice might prevent the FBI from “finding the truth.” But this decision poses a huge legitimacy risk to the bureau, because the failure to notify can later be construed as evidence of political motivation. The FBI needs a clear policy and procedures on when and how it informs presidential campaigns about investigations.
Horowitz identified other important gaps in the standing FBI guidance about investigations related to political campaigns. As he set out in his report and his testimony to the Senate, the use of confidential human sources for “consensually monitoring” individuals affiliated with the Trump campaign, including an official who was not a subject of the investigation, was “permitted under Department and FBI policy because their use was not for the sole purpose of monitoring activities protected by the First Amendment.” However, he noted that the FBI rules allow agents to seek approval of this kind of intrusive investigative technique from only a first-level supervisor, without any requirement that the FBI notify senior Justice Department officials. Horowitz recommended a requirement that there be “consultation, at a minimum.” The department should consider a rule requiring both notice and approval in cases involving politicians and campaigns.
Last week is not the first time that Horowitz has reported on gaps in Justice Department rules and policies in this context. In his 2018 report on the Clinton email investigation, he noted that departmental memoranda counseled that “politics must play no role in the decisions of federal investigators or prosecutors regarding any investigations or criminal charges.” However, enforcement of this policy has rested on mere norms or practices against overt investigative steps within two months of a primary or general election, rather than something firmer and more concrete. Horowitz recommended that the department “consider” additional “guidance to agents and prosecutors concerning the taking of overt investigative steps, indictments, public announcements, or other actions that could impact an election.” The department must establish clear policies on these matters.
A second major problem identified by Horowitz concerns the 1978 Foreign Intelligence Surveillance Act (FISA) for obtaining judicial approval for electronic surveillance of foreign agents. This was the process under which the Justice Department got approval four times to surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review a proposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the Justice Department and FBI, were held up as reasons for confidence in the process.
The Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts related to the surveillance. The system lacks an adversarial process; no one represents the interests of the target. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court. All of the inaccuracies, omissions, or errors concerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page.
Horowitz concluded that these matters were known inside the FBI but not conveyed to relevant officials in the Justice Department. He described these problems as “serious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications.” He expressed dismay that “so many basic and fundamental errors were made by [FBI agents] on one of the most sensitive FBI investigations that was briefed to the highest levels within the FBI, and that FBI officials expected would eventually be subjected to close scrutiny.” And he concluded that the episode “raised significant questions regarding the FBI chain of command's management and supervision of the FISA process.”
When the political dust settles, if not sooner, this aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny after the fact, the process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.
Horowitz recommended a number of reforms to ensure factual accuracy and completeness, and FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to consider adding criminal or civil penalties for gross factual misrepresentations before the FISA court, akin to the penalties in FISA for electronic surveillance without court approval. It should also consider ways to introduce some sort of adversarial process, at least in the most sensitive contexts. | Horowitz’s report
told a complex story
Its publication predictably ignited a contest between Dem
s and Rep
s to extract
what was most serviceable in the construction of competing political “narratives.”
But there is something
more important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed
suggesting
an opportunity
if Congress
can seize it
The mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party
or to serve the FBI’s institutional interests at the expense of elected officials.
These tactics, when publicly disclosed in the
70s, almost destroyed the bureau. The FBI survived with its
powers intact only because it entered
a “grand bargain” with Congress
about the use of such powers
The central elements of that bargain
were
The
intel
bureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions
The idea was that the use of these powers would be
watched by institutions that could be trusted
The point of these arrangements was
to assure the American people that the FBI (and other intelligence agencies) did not in fact use
powers for political ends, but also to ensure that it did not appear to use its powers for these ends.
The FBI has taken a large hit in its credibility
due in large part to Trump’s
attacks on it. But the bureau has also hurt itself
Horowitz
documented
serious failures of policy, process, and decision making
While Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.
The problems
do not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive
The overall effect of these controversies has been to call into question the viability of the grand bargain
especially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But
areas warrant special consideration
A
major problem identified by Horowitz concerns
FISA
This was the process under which the Justice Department got approval
to surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review
proposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the
FBI, were held up as reasons for confidence in the process
The Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts
The system lacks an adversarial process; no one represents the
target. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court
All
concerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page
Horowitz concluded that these matters were known inside the FBI but not conveyed
serious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications
When the political dust settles
this aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny
the process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.
Horowitz recommended a number of reforms
and FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to
introduce some sort of adversarial process | 239 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.”",
"role": "system"
},
{
"content": "Inspector General Michael Horowitz’s report last week told a complex story about extraordinary events related to the investigation of officials in Donald Trump’s 2016 presidential campaign. Its publication predictably ignited a contest between Democrats and Republicans to extract from the 429-page opus what was most serviceable in the construction of competing political “narratives.”\nBut there is something much more important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed in how the FBI and the Department of Justice, of which it is a component, open and conduct investigations—especially those related to politicians and political campaigns. The report prompted concerns from both sides of the aisle, suggesting that there’s an opportunity for serious reflection and reform—if Congress and the executive branch can seize it.\nThe mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party over another, or to serve the FBI’s institutional interests at the expense of elected officials.\nThese tactics, when publicly disclosed in the 1960s and ’70s, almost destroyed the bureau. The FBI survived with its powerful investigatory and surveillance powers intact only because it entered into a “grand bargain” with Congress and the American people about the use of such powers.\nThe central elements of that bargain, as one of us previously explained, were as follows:\nThe president and his intelligence bureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions on how they collected, analyzed, and disseminated intelligence information; a bevy of lawyers throughout the intelligence community and, over time, in the Justice Department monitored and enforced those restrictions; domestic surveillance required a court order, including a court order from a new court, the Foreign Intelligence Surveillance Court, for foreign intelligence investigations; and two new committees, the Senate and House Intelligence committees, were to be kept “fully and currently informed” of all significant intelligence activities, and would have robust oversight authorities. The idea was that the use of these powers would be documented and watched by institutions that could be trusted to keep secrets but would act as credible surrogates for public oversight mechanisms.\nThe point of these arrangements was not only to assure the American people that the FBI (and other intelligence agencies) did not in fact use its powers for political ends, but also to ensure that it did not appear to use its powers for these ends.\nThe FBI has taken a large hit in its credibility over the past four years, due in large part to Trump’s unprecedented, reckless, and routinely baseless attacks on it. But the bureau has also hurt itself through its conduct of the investigation of Trump-campaign officials and of Hillary Clinton’s emails when she was a presidential candidate. Horowitz, in his latest report, and in three earlier ones related to the 2016 election, has documented what appear to be serious failures of policy, process, and decision making at multiple levels. \nWhile Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee. Among Democrats’ numerous other concerns about the conduct of the Clinton email investigation, they were appalled by reasons to believe that a “a deep and visceral hatred” of Hillary Clinton in the FBI’s New York field office led agents to leak investigative information to Rudy Giuliani and Devin Nunes, and that the threat of further leaks informed the FBI’s decision to notify Congress, only 11 days before the 2016 election, that it had reopened the Clinton email investigation. \nThe problems at the FBI revealed by multiple inspector-general investigations do not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive or manipulation. The overall effect of these controversies has been to call into question the viability of the grand bargain, at least on its current terms, especially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But based on his insights and other factors, we think that three areas warrant special consideration. \nFirst is the serious problem of inadequate guidance to the FBI for opening and conducting investigations against politicians or campaigns, especially during election season. The FBI has two guidebooks—the Attorney General’s Guidelines, and the Domestic Investigations and Operations Guide (DIOG)—to regulate the bureau’s investigative and intelligence-collection activities. Such guidance is vital to FBI legitimacy, especially when it investigates a politician or campaign. Well-designed guidance can establish a process to ensure that controversial steps are properly vetted in ways that minimize mistakes or disproportionate action. This, in turn, can help the FBI avoid controversy later, either because it declined to take unnecessarily controversial investigatory steps, or because it took those steps but can point to clear guidance for legitimation. But the past three years have shown that the current guidance is not well designed for investigations related to political campaigns. \nAs Horowitz noted, the threshold for opening an FBI investigation is “low.” To open a “full investigation” like the one against Trump-campaign officials—which allows the FBI to use all its prodigious surveillance and intelligence-gathering techniques—the FBI needs to satisfy only two requirements. First, it must have an “authorized purpose” (such as detecting a federal crime or threat to national security). And second, it must have an “articulable factual basis” that “reasonably indicates” that a federal crime or threat to national security “may have occurred” or “may occur,” and that the investigation may obtain relevant information. These relatively permissive criteria give officials enormous leeway in opening an investigation based merely on the right type of factual allegations. And there is only one additional rule when the investigation involves a political candidate or member of a political organization: The FBI agent considering such a step must consult up the FBI chain of command to various degrees depending on the type of assessment or investigation in issue. \nHorowitz’s latest report expressed surprise about deficiencies in the DIOG guidance in the context of opening investigations related to political campaigns. He proposed that the FBI establish procedures that require giving notice not just to a senior FBI official, but also outside the bureau to a senior DOJ official, such as the deputy attorney general, “for case openings that implicate core First Amendment activity.” Such elevated notice at the outset of such a sensitive investigation is a good idea because it brings into the loop a Justice Department official who has been subject to Senate confirmation and is directly subject to congressional oversight, and thus politically accountable. Such high-level notice occurs in other crucial circumstances, such as sensitive undercover operations. A senior Senate-confirmed Justice Department official will also likely take a broader view of the matter than the Bureau and can, as Horowitz noted in his hearing, “consider the potential constitutional and prudential implications in advance of these activities.” In the event that such an official fails to do so, or makes the attempt but exercises poor judgment, he or she is the one who should answer for this to Congress and public. \nWe would go further and propose not just better notice rules up the chain of command, but also concrete guidance to FBI officials about when they can or cannot open cases in politically sensitive areas. Horowitz noted that the FBI rulebook uses the same low threshold for opening sensitive cases connected to a political campaign, which implicate First Amendment activity and electoral integrity, as in ordinary, nonsensitive cases. The FBI should develop a heightened threshold for opening such cases to ensure that First Amendment activity is not unduly investigated. It is a fair concern that the new standard could invite risk aversion within the bureau and leave undetected criminal or adverse national-security activity related to a campaign. This requires particularly close attention in an era likely to be marked by increased activity to illicitly influence our political process. But given the very large costs of the investigations in the past three years, these risks seem acceptable and can be mitigated if the heightened standard is carefully crafted. \nWe would also propose a heightened standard, and mandatory notice to the Justice Department, before the FBI opens a counterintelligence investigation on the president, as it did in May 2017. Such investigations focus on threats to national security from foreign intelligence agencies. The FBI is authorized to open such investigations but lacks specific guidance when the possible foreign asset is the president. The FBI appears to have concluded that Trump posed a national-security threat, and warranted counterintelligence scrutiny, because he was harming the FBI’s Russia investigation. But the president is not like other individuals who might pose a threat to national security, because he is constitutionally empowered to determine the national-security interests of the United States and to conduct its foreign policy, which includes the authority to change foreign policy on a dime. \nWe are not saying that the FBI should be barred from investigating a president who appears to be a Manchurian candidate, but the situation requires special guidance. The lines between a Manchurian candidate, a president engaged in quid pro quo foreign-policy bargaining that is possibly corrupt, and the very large discretionary control that presidents have to conduct U.S. foreign policy are not always easy to discern. The decision to open a counterintelligence investigation of the president is so consequential for the nation (and the FBI), and so politically fraught, that it should not be made without much clearer guidance. It also should not be made by the FBI alone. FBI leadership wisely realized this truth, even without specific guidance, because they quickly briefed the acting attorney general and relevant congressional leaders. Such briefings should be mandatory—both for the FBI’s sake, and to ensure that Justice Department leaders do not eschew their responsibilities. \nHorowitz additionally noted that the FBI had no policy on whether to give notice to campaigns, like Trump’s, that are subject to investigation. According to Horowitz, the discretionary call not to inform the Trump campaign was made by an assistant director of an FBI division, who concluded without any guidance that such notice might prevent the FBI from “finding the truth.” But this decision poses a huge legitimacy risk to the bureau, because the failure to notify can later be construed as evidence of political motivation. The FBI needs a clear policy and procedures on when and how it informs presidential campaigns about investigations. \nHorowitz identified other important gaps in the standing FBI guidance about investigations related to political campaigns. As he set out in his report and his testimony to the Senate, the use of confidential human sources for “consensually monitoring” individuals affiliated with the Trump campaign, including an official who was not a subject of the investigation, was “permitted under Department and FBI policy because their use was not for the sole purpose of monitoring activities protected by the First Amendment.” However, he noted that the FBI rules allow agents to seek approval of this kind of intrusive investigative technique from only a first-level supervisor, without any requirement that the FBI notify senior Justice Department officials. Horowitz recommended a requirement that there be “consultation, at a minimum.” The department should consider a rule requiring both notice and approval in cases involving politicians and campaigns. \nLast week is not the first time that Horowitz has reported on gaps in Justice Department rules and policies in this context. In his 2018 report on the Clinton email investigation, he noted that departmental memoranda counseled that “politics must play no role in the decisions of federal investigators or prosecutors regarding any investigations or criminal charges.” However, enforcement of this policy has rested on mere norms or practices against overt investigative steps within two months of a primary or general election, rather than something firmer and more concrete. Horowitz recommended that the department “consider” additional “guidance to agents and prosecutors concerning the taking of overt investigative steps, indictments, public announcements, or other actions that could impact an election.” The department must establish clear policies on these matters. \nA second major problem identified by Horowitz concerns the 1978 Foreign Intelligence Surveillance Act (FISA) for obtaining judicial approval for electronic surveillance of foreign agents. This was the process under which the Justice Department got approval four times to surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review a proposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the Justice Department and FBI, were held up as reasons for confidence in the process. \nThe Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts related to the surveillance. The system lacks an adversarial process; no one represents the interests of the target. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court. All of the inaccuracies, omissions, or errors concerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page. \nHorowitz concluded that these matters were known inside the FBI but not conveyed to relevant officials in the Justice Department. He described these problems as “serious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications.” He expressed dismay that “so many basic and fundamental errors were made by [FBI agents] on one of the most sensitive FBI investigations that was briefed to the highest levels within the FBI, and that FBI officials expected would eventually be subjected to close scrutiny.” And he concluded that the episode “raised significant questions regarding the FBI chain of command's management and supervision of the FISA process.” \nWhen the political dust settles, if not sooner, this aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny after the fact, the process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases. \nHorowitz recommended a number of reforms to ensure factual accuracy and completeness, and FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to consider adding criminal or civil penalties for gross factual misrepresentations before the FISA court, akin to the penalties in FISA for electronic surveillance without court approval. It should also consider ways to introduce some sort of adversarial process, at least in the most sensitive contexts. ",
"role": "user"
},
{
"content": "Horowitz’s report\ntold a complex story\nIts publication predictably ignited a contest between Dem\ns and Rep\ns to extract\nwhat was most serviceable in the construction of competing political “narratives.”\nBut there is something\nmore important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed\nsuggesting\nan opportunity\nif Congress\ncan seize it\nThe mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party\nor to serve the FBI’s institutional interests at the expense of elected officials.\nThese tactics, when publicly disclosed in the\n70s, almost destroyed the bureau. The FBI survived with its\npowers intact only because it entered\na “grand bargain” with Congress\nabout the use of such powers\nThe central elements of that bargain\nwere\nThe\nintel\nbureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions\nThe idea was that the use of these powers would be\nwatched by institutions that could be trusted\nThe point of these arrangements was\nto assure the American people that the FBI (and other intelligence agencies) did not in fact use\npowers for political ends, but also to ensure that it did not appear to use its powers for these ends.\nThe FBI has taken a large hit in its credibility\ndue in large part to Trump’s\nattacks on it. But the bureau has also hurt itself\nHorowitz\ndocumented\nserious failures of policy, process, and decision making\nWhile Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.\nThe problems\ndo not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive\nThe overall effect of these controversies has been to call into question the viability of the grand bargain\nespecially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But\nareas warrant special consideration\nA\nmajor problem identified by Horowitz concerns\nFISA\nThis was the process under which the Justice Department got approval\nto surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review\nproposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the\nFBI, were held up as reasons for confidence in the process\nThe Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts\nThe system lacks an adversarial process; no one represents the\ntarget. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court\nAll\nconcerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page\nHorowitz concluded that these matters were known inside the FBI but not conveyed\nserious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications\nWhen the political dust settles\nthis aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny\nthe process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.\nHorowitz recommended a number of reforms\nand FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to\nintroduce some sort of adversarial process",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.” | Horowitz’s report
told a complex story
Its publication predictably ignited a contest between Dem
s and Rep
s to extract
what was most serviceable in the construction of competing political “narratives.”
But there is something
more important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed
suggesting
an opportunity
if Congress
can seize it
The mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party
or to serve the FBI’s institutional interests at the expense of elected officials.
These tactics, when publicly disclosed in the
70s, almost destroyed the bureau. The FBI survived with its
powers intact only because it entered
a “grand bargain” with Congress
about the use of such powers
The central elements of that bargain
were
The
intel
bureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions
The idea was that the use of these powers would be
watched by institutions that could be trusted
The point of these arrangements was
to assure the American people that the FBI (and other intelligence agencies) did not in fact use
powers for political ends, but also to ensure that it did not appear to use its powers for these ends.
The FBI has taken a large hit in its credibility
due in large part to Trump’s
attacks on it. But the bureau has also hurt itself
Horowitz
documented
serious failures of policy, process, and decision making
While Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.
The problems
do not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive
The overall effect of these controversies has been to call into question the viability of the grand bargain
especially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But
areas warrant special consideration
A
major problem identified by Horowitz concerns
FISA
This was the process under which the Justice Department got approval
to surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review
proposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the
FBI, were held up as reasons for confidence in the process
The Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts
The system lacks an adversarial process; no one represents the
target. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court
All
concerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page
Horowitz concluded that these matters were known inside the FBI but not conveyed
serious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications
When the political dust settles
this aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny
the process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.
Horowitz recommended a number of reforms
and FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to
introduce some sort of adversarial process | Horowitz
contest between Dem
s
Rep
s
extract
most serviceable in the construction of competing political “narratives.”
more important
reforms are vital
opportunity
Congress
seize it
mere
hint
protected speech
political activity
taint the democratic process
influencing electoral outcomes
long history
politic
s
including
one party
publicly disclosed
70s
destroyed the bureau
survived
because it entered
a “grand bargain”
Congress
central elements
intel
maintain robust surveillance
espionage
including domestically
in exchange
Congress subject
significant legal restrictions
watched
institutions
trusted
point
in fact
politic
s
appear
large hit in its credibility
Trump’s
attacks
also hurt itself
Horowitz
serious failures
policy, process, and decision making
turned on different interests at different times
fairly
human error
flawed judgment
not to political machinations
little chance of being heard clearly or accepted widely in the heat of partisan conflict
charg
political motive
overall effect
call into question the viability of the grand bargain
special consideration
major problem
FISA
Carter Page
core element of the grand bargain
main innovation
review
decades
inside
confidence
Achilles’ heel
depend
accurate
representations
FBI
full accounting of the facts
adversarial process
target
most troubling
All
undermine the justification
inside
not conveyed
serious performance failures
political dust settles
crisis for the entire FISA process
rests
accuracy
FBI
facts
adverse
most serious of possible contexts
highest
everyone knew there would be close scrutiny
broke
altogether
ordinary
number of reforms
not clear that this problem can be fixed credibly within the executive branch alone
Congress
adversarial process | 240 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.”",
"role": "system"
},
{
"content": "Horowitz’s report\ntold a complex story\nIts publication predictably ignited a contest between Dem\ns and Rep\ns to extract\nwhat was most serviceable in the construction of competing political “narratives.”\nBut there is something\nmore important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed\nsuggesting\nan opportunity\nif Congress\ncan seize it\nThe mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party\nor to serve the FBI’s institutional interests at the expense of elected officials.\nThese tactics, when publicly disclosed in the\n70s, almost destroyed the bureau. The FBI survived with its\npowers intact only because it entered\na “grand bargain” with Congress\nabout the use of such powers\nThe central elements of that bargain\nwere\nThe\nintel\nbureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions\nThe idea was that the use of these powers would be\nwatched by institutions that could be trusted\nThe point of these arrangements was\nto assure the American people that the FBI (and other intelligence agencies) did not in fact use\npowers for political ends, but also to ensure that it did not appear to use its powers for these ends.\nThe FBI has taken a large hit in its credibility\ndue in large part to Trump’s\nattacks on it. But the bureau has also hurt itself\nHorowitz\ndocumented\nserious failures of policy, process, and decision making\nWhile Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.\nThe problems\ndo not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive\nThe overall effect of these controversies has been to call into question the viability of the grand bargain\nespecially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But\nareas warrant special consideration\nA\nmajor problem identified by Horowitz concerns\nFISA\nThis was the process under which the Justice Department got approval\nto surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review\nproposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the\nFBI, were held up as reasons for confidence in the process\nThe Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts\nThe system lacks an adversarial process; no one represents the\ntarget. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court\nAll\nconcerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page\nHorowitz concluded that these matters were known inside the FBI but not conveyed\nserious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications\nWhen the political dust settles\nthis aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny\nthe process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.\nHorowitz recommended a number of reforms\nand FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to\nintroduce some sort of adversarial process",
"role": "user"
},
{
"content": "Horowitz\ncontest between Dem\ns\nRep\ns\nextract\nmost serviceable in the construction of competing political “narratives.”\nmore important\nreforms are vital\nopportunity\nCongress\nseize it\nmere\nhint\nprotected speech\npolitical activity\ntaint the democratic process\ninfluencing electoral outcomes\nlong history\npolitic\ns\nincluding\none party\npublicly disclosed\n70s\ndestroyed the bureau\nsurvived\nbecause it entered\na “grand bargain”\nCongress\ncentral elements\nintel\nmaintain robust surveillance\nespionage\nincluding domestically\nin exchange\nCongress subject\nsignificant legal restrictions\nwatched\ninstitutions\ntrusted\npoint\nin fact\npolitic\ns\nappear\nlarge hit in its credibility\nTrump’s\nattacks\nalso hurt itself\nHorowitz\nserious failures\npolicy, process, and decision making\nturned on different interests at different times\nfairly\nhuman error\nflawed judgment\nnot to political machinations\nlittle chance of being heard clearly or accepted widely in the heat of partisan conflict\ncharg\npolitical motive\noverall effect\ncall into question the viability of the grand bargain\nspecial consideration\nmajor problem\nFISA\nCarter Page\ncore element of the grand bargain\nmain innovation\nreview\ndecades\ninside\nconfidence\nAchilles’ heel\ndepend\naccurate\nrepresentations\nFBI\nfull accounting of the facts\nadversarial process\ntarget\nmost troubling\nAll\nundermine the justification\ninside\nnot conveyed\nserious performance failures\npolitical dust settles\ncrisis for the entire FISA process\nrests\naccuracy\nFBI\nfacts\nadverse\nmost serious of possible contexts\nhighest\neveryone knew there would be close scrutiny\nbroke\naltogether\nordinary\nnumber of reforms\nnot clear that this problem can be fixed credibly within the executive branch alone\nCongress\nadversarial process",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.”
Highlight Ratio: 0.2 | Horowitz’s report
told a complex story
Its publication predictably ignited a contest between Dem
s and Rep
s to extract
what was most serviceable in the construction of competing political “narratives.”
But there is something
more important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed
suggesting
an opportunity
if Congress
can seize it
The mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party
or to serve the FBI’s institutional interests at the expense of elected officials.
These tactics, when publicly disclosed in the
70s, almost destroyed the bureau. The FBI survived with its
powers intact only because it entered
a “grand bargain” with Congress
about the use of such powers
The central elements of that bargain
were
The
intel
bureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions
The idea was that the use of these powers would be
watched by institutions that could be trusted
The point of these arrangements was
to assure the American people that the FBI (and other intelligence agencies) did not in fact use
powers for political ends, but also to ensure that it did not appear to use its powers for these ends.
The FBI has taken a large hit in its credibility
due in large part to Trump’s
attacks on it. But the bureau has also hurt itself
Horowitz
documented
serious failures of policy, process, and decision making
While Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.
The problems
do not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive
The overall effect of these controversies has been to call into question the viability of the grand bargain
especially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But
areas warrant special consideration
A
major problem identified by Horowitz concerns
FISA
This was the process under which the Justice Department got approval
to surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review
proposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the
FBI, were held up as reasons for confidence in the process
The Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts
The system lacks an adversarial process; no one represents the
target. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court
All
concerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page
Horowitz concluded that these matters were known inside the FBI but not conveyed
serious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications
When the political dust settles
this aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny
the process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.
Horowitz recommended a number of reforms
and FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to
introduce some sort of adversarial process | the Horowitz report
shows
reforms are vital
an opportunity
if Congress
can seize it
The mere
hint of investigation into a campaign
threatens
democratic process
the FBI
history of
surveillance
for politic
s
when
disclosed in the
70s, almost destroyed the bureau
survived
because it entered
a “grand bargain” with Congress
intel
maintain
surveillance
in exchange, Congress subject
them [to]
legal restrictions
to assure
that
agencies) did not in fact use
powers for politic
s
also
that it did not appear to
even where
breakdown can fairly be attributed to human error
this explanation stands little chance
in the heat of partisan conflict. One side
will
charg
a political motive
The
effect
call into question
the grand bargain
FISA is a core
of the grand bargain
The Achilles’ heel
is that
judges
depend
on accurate
representations by
the FBI
lacks
adversarial process
When
political dust settles
this
will spark
crisis for the entire FISA process. The system rests on
accuracy
of
FBI
facts
when the stakes were highest
the process
broke
altogether
it is not clear
this problem can be fixed credibly within the executive
alone. Congress needs to
introduce
adversarial process | 241 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: The recent FBI Inspector General report revealed systematic process failures in warrant applications for foreign intelligence gathering---also called “FISA warrants”---stemming from a lack of external oversight. This pattern of abuses politicizes intel-gathering---undermining the broader legitimacy of the intelligence community, or “IC.”\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "Horowitz’s report\ntold a complex story\nIts publication predictably ignited a contest between Dem\ns and Rep\ns to extract\nwhat was most serviceable in the construction of competing political “narratives.”\nBut there is something\nmore important in the Horowitz report than evidence for political vindication. The report shows that serious reforms are vitally needed\nsuggesting\nan opportunity\nif Congress\ncan seize it\nThe mere fact or even hint of investigation into a campaign threatens protected speech and political activity, and, worse yet, threatens to taint the democratic process by unduly influencing electoral outcomes. Such investigations are all the more fraught because the FBI has a long history of using investigations and secret surveillance against American citizens for political ends, including to favor one party\nor to serve the FBI’s institutional interests at the expense of elected officials.\nThese tactics, when publicly disclosed in the\n70s, almost destroyed the bureau. The FBI survived with its\npowers intact only because it entered\na “grand bargain” with Congress\nabout the use of such powers\nThe central elements of that bargain\nwere\nThe\nintel\nbureaucracy were allowed to maintain robust surveillance and espionage capacities, including domestically. But in exchange, Congress subjected them [to] significant legal restrictions\nThe idea was that the use of these powers would be\nwatched by institutions that could be trusted\nThe point of these arrangements was\nto assure the American people that the FBI (and other intelligence agencies) did not in fact use\npowers for political ends, but also to ensure that it did not appear to use its powers for these ends.\nThe FBI has taken a large hit in its credibility\ndue in large part to Trump’s\nattacks on it. But the bureau has also hurt itself\nHorowitz\ndocumented\nserious failures of policy, process, and decision making\nWhile Republicans now link the failings detailed by Horowitz to a “deep state” that is biased against their party and determined to take down Trump, Democrats remain aggrieved by bureau actions that they believe fatally damaged their 2016 presidential nominee.\nThe problems\ndo not cut politically in one direction. To the extent that an FBI operating without clear guidance and rigorous process can be weaponized to serve partisan political interests, the weapons are available to be turned on different interests at different times. Moreover, even where systemic breakdown can fairly be attributed to human error and flawed judgment, not to political machinations, this explanation stands little chance of being heard clearly or accepted widely in the heat of partisan conflict. One side or the other will find in this kind of breakdown the basis for charging a political motive\nThe overall effect of these controversies has been to call into question the viability of the grand bargain\nespecially as it concerns FBI action related to politicians and political campaigns. Horowitz’s reports provide numerous fine-grained and technical criticisms, and literally dozens of proposals for reform, to enhance FBI process, credibility, and legitimacy in this context. They all deserve attention. But\nareas warrant special consideration\nA\nmajor problem identified by Horowitz concerns\nFISA\nThis was the process under which the Justice Department got approval\nto surveil Carter Page. FISA is a core element of the grand bargain. The main innovation in FISA, and the main basis for its legitimation, is that federal judges review\nproposed surveillance and approve it only if the government can show probable cause that the target of the surveillance is an “agent of a foreign power.” For decades the involvement of independent judges, combined with the supposedly rigorous procedures inside the\nFBI, were held up as reasons for confidence in the process\nThe Achilles’ heel in the FISA process is that the judges are dependent on accurate and full factual representations by the Justice Department, which in turn is dependent on the FBI for a full accounting of the facts\nThe system lacks an adversarial process; no one represents the\ntarget. Here is where Horowitz’s report is most troubling. He found 17 significant inaccuracies, omissions, or errors related to the four applications to surveil Page, many (but not all) of which resulted in incomplete or misleading information being provided to the FISA court\nAll\nconcerned matters that, if handled properly, would have tended to undermine the justification for surveilling Page\nHorowitz concluded that these matters were known inside the FBI but not conveyed\nserious performance failures by the supervisory and non-supervisory [FBI] agents with responsibility over the FISA applications\nWhen the political dust settles\nthis aspect of the Horowitz report will spark a crisis for the entire FISA process. The system rests on the accuracy and fullness of the FBI’s factual representations, including facts that are adverse to the surveillance. In the most serious of possible contexts, when the stakes were highest and everyone knew there would be close scrutiny\nthe process appears to have broken down altogether—raising the question of whether the system can be trusted in ordinary cases.\nHorowitz recommended a number of reforms\nand FBI Director Christopher Wray is taking these proposals very seriously. But it is not clear that this problem can be fixed credibly within the executive branch alone. Congress needs to\nintroduce some sort of adversarial process",
"role": "user"
},
{
"content": "the Horowitz report\nshows\nreforms are vital\nan opportunity\nif Congress\ncan seize it\nThe mere\nhint of investigation into a campaign\nthreatens\ndemocratic process\nthe FBI\nhistory of\nsurveillance\nfor politic\ns\nwhen\ndisclosed in the\n70s, almost destroyed the bureau\nsurvived\nbecause it entered\na “grand bargain” with Congress\nintel\nmaintain\nsurveillance\nin exchange, Congress subject\nthem [to]\nlegal restrictions\nto assure\nthat\nagencies) did not in fact use\npowers for politic\ns\nalso\nthat it did not appear to\neven where\nbreakdown can fairly be attributed to human error\nthis explanation stands little chance\nin the heat of partisan conflict. One side\nwill\ncharg\na political motive\nThe\neffect\ncall into question\nthe grand bargain\nFISA is a core\nof the grand bargain\nThe Achilles’ heel\nis that\njudges\ndepend\non accurate\nrepresentations by\nthe FBI\nlacks\nadversarial process\nWhen\npolitical dust settles\nthis\nwill spark\ncrisis for the entire FISA process. The system rests on\naccuracy\nof\nFBI\nfacts\nwhen the stakes were highest\nthe process\nbroke\naltogether\nit is not clear\nthis problem can be fixed credibly within the executive\nalone. Congress needs to\nintroduce\nadversarial process",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks | There is now zero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire intelligence community to both Presidents Obama and Trump. 1 Reasonable differences of informed opinion remain about whether these were successful enough to swing the election from Hillary Clinton to Mr. Trump, but they probably were according to Kathleen Hall Jamieson, Director of the Annenberg Public Policy Center at the University of Pennsylvania. 2 Obviously, if Russian efforts were enough to swing an American Presidential election this would be a profound political effect at minimal cost in national security expenditures (for the Russians anyway). But I want to focus for a moment on some other meta-goals of the Russian effort that were unambiguously successful, regardless of whether they tipped the election to Mr. Trump.
a. Doubt in the validity of electoral process in America in general was greatly amplified.
b. Doubt in the legitimacy of many basic institutions, like of our intelligence community, law enforcement, media, the Department of Justice and other departments was increased.
c. Polarization of politics in America, already severe, was also increased, and
d. Hatred among various American communities was increased. Sewing division, uncertainty and illegitimacy were arguably among Russia’s main strategic goals.
e. Relationships with long-term allies were put under great stress. NATO suffered. This was another Russian goal. Fractured US relations with Canada and Mexico were bonuses.
f. The term “Fake News” became a meme, with pernicious effects far beyond the obvious. Democratic intelligence systems depend on concepts like evidence-based analysis and on objective presentation to policy makers of “truth” as best that can be discerned. When these fundamental concepts are shattered, trust evaporates, and hell can result.
Special Counsel Robert Mueller’s report on these topics has not been released to the American public at the time of writing this paper, but it will probably have significant effects on both academic conclusions and on practical politics in America. I can still conclude with high confidence now that this was one of the most effective applications in recent history of what we call “psychological operations” and what Russians often call “active measures.” 3
The general dilemma of election meddling with practical consequences for neighbors is nothing new to Eastern Europe. They have dealt with Russian (and other) meddling for a very long time. America is just now learning the hard way why so many people in so many countries hate us for past acts when we manipulated their elections. That negative reaction is a vivid example of blowback from an intelligence practitioner’s or policy maker’s point of view. Policy makers chose a somewhat evil method (manipulating a neighbors’ elections) to support some allegedly good end (like containing communism) but found that in the long run, this also creates much hatred of your country. Hatred has national security consequences too. The internet evolution to distributed networks of information flow, and the special power of “social media” to mold public perceptions of reality are technical aspects of this problem, but the moral dilemma is eternal.
Excellent technical reviews of evolving methods of manipulation through social media and other emerging fora can be found by Matteo Bonfanti of Zurich’s ETS, and James Pamment et al. of Lund University and Sweden’s Civil Contingencies Agency (MSB). 4 They describe new techniques so exotic they require a new language, as noted in this quote from Bonfanti’s article. “They include socio-cognitive and psychographic hacking via dark advertisement, social hacking, band wagoning, digital disinformation and fakes, exploitation via bots, botnets and sock-puppets, trolling and flaming.” 5 Greg Treverton has also written about these challenges for the Swedish Defense University. 6 Finally, the Canadian Security Intelligence Service published a very good workshop overview on such techniques, but as is their practice all papers are anonymous. 7 By tapping into and enflaming local prejudices, the Russians spread lies faster and further than truth* broadcast by establishment, mainstream media. 8
Should we promote truth and goodness in our public diplomacy and private (even secret) actions, or descend to deception and evil in order to gain advantage in our competitions for power, wealth and basic security? Well in war, this seems simple. If you do not win, you and all you love may die young. But in cold wars or normal competitions, this is not simple. Evil methods tend to yield evil outcomes far more often than not. Unintended consequences abound, and are often negative. A similar dilemma occurs when people consider assassination (or “targeted killings”) as they have throughout the centuries and to this day. A particular and interesting contemporary case lies at the intersection of these phenomena, the much-publicized murder of Saudi journalist Jamal Khashoggi at the Saudi Arabian consulate in Istanbul, Turkey, on October 2, 2018. 9
Assassinations of Journalists and other Suppressions of “Objective Reality”
“Fake News” can be created in very large volumes by active propaganda efforts funded by states. This can be enhanced by suppression of alternative “narratives.” One very harsh method for that is the murder of journalists who do not comply with official stories. 10 There are many other methods short of murder, like arresting, intimidating or buying off large numbers of independent or critical journalists. And police-states simply own all significant media and fire any journalist who strays from party lines, as occurs in North Korea, China and other police-states these days.
Many more countries than Saudi Arabia use such tactics. Some are not police-states. For one sad example, NATO ally Turkey, which exposed the especially gruesome and egregious murder of Jamal Khashoggi, imprisons more journalists today than any other nation on earth (as best we can measure such tragedies) due to President Erdoğan’s campaign against supporters of Fetullah Gulan. 11 , 12 Those include many journalists, professors and whole media outlets and schools in Turkey (many since closed down by the government). Turkey is not yet a police-state, although it trends in that direction in 2019. Venezuela provides another example from another continent, as Nicolás Maduro drives his formerly prosperous country towards failed-state conditions. In most such cases, control of the media is a prelude to rot within, but both Russia and China prove that police-states with controlled media can endure for very long periods.
True police-states have always been in the vanguard of these kinds of efforts to control people’s thoughts by controlling the information available to them. But even undeniable and relatively healthy democracies can be guilty of egregious controls on information, including kidnapping journalists and whistleblowers, imprisonment of critics and even assassinations of people who tell inconvenient truths. The saddest examples of all to me come from Israel since it once aspired to be a leader in law and moral thought. Then they kidnapped Israeli citizen Mordecai Vanunu overseas because he told the London Sunday Times about Israel’s clandestine nuclear weapons program. 13 They then held Vanunu in solitary confinement for years lest he share more with international media. After 18 years and release, they convicted him again, for speaking to foreigners about his plight. Meanwhile, every child in neighbor nations knows they could be burned alive by Israel’s nuclear arsenal. Israel has also assassinated many hundreds, probably thousands of its neighbors in Lebanon, Egypt, Syria and Iran. Those include both Palestinians working for their own freedom (and sometimes fighting for their freedom, of course) and nationals of those neighbor states, for many reasons detailed in the incomparable record of Israeli assassinations called “Rise and Kill First,” a book reviewed and approved by IDF military censors in 2018. 14 This source is unique, because it includes the moral dilemmas and angst faced by top Israeli intelligence officials as they decided to kill many, many hundreds of possible enemies of their state. Despite strenuous efforts to avoid injuring bystanders, innocents and witnesses nearby were also killed more often than anyone wants to admit. Hence the intense secrecy about “targeted killing” operations, even in a state that aspires to moral leadership.
I will not detail assassinations attempted, conducted or enabled by my own United States of America because it continues to deny such things. The official denials are lies, but they are important to keep practitioners loyal to my country that also once aspired to moral leadership.
It appears that no nation is immune to the temptation to use unconventional methods to attack people or institutions that appear to be threats to the state. “Unconventional methods” include here (more or less) precise targeted killings and massive information operations against whole populations that exploit every dimension of the internet including social media in particular. What happens to strategic intelligence in this context of constant battle against endless enemies? The short answer is that strategic intelligence is commonly ignored, so the enemies never end.
Nuclear Weapons, other WMD, the Developing Global Crisis, and Strategic Intelligence
Strategic intelligence is often ignored simply because most of the system is very busy every day with current intelligence about active armed conflicts and responses to the blizzard of modern propaganda operations. Already a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings, far too long-term, a purview of mostly elder analysts on end of career assignments, and is almost never attended to by active policy makers.
I will illustrate this by reference to something virtually everyone agrees is a real threat to all of human civilization, nuclear and other WMD, which proliferate as we chat about terrorists. I will also describe very briefly a “Developing Global Crisis” that underlies much of the lethal conflict on earth today. Few policy makers or analysts will dispute that each of those causes of threat to everything under heaven are real, and powerful. But little is done about any of those ultimate causes of armed conflicts, because they are very difficult to deal with, and some are protected by “taboos” (like population pressure). Finally, everyone is so busy fighting the perpetual wars they are trapped in. So the wars continue. Intelligence professionals can describe in detail why they so often ignore ultimate threats to focus on the mosquitos that are biting states today.
Many factors contribute to the “Developing Global Crisis” so descriptions can be cumbersome, but all include at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law. 15 These are powerful causes of wars over long periods of history. They are especially relevant today as we become ever more global, economically and in many other ways. But they typically underlie more obvious proximate causes of each particular conflict. This is especially true of the first two. Everyone knows that competition for resources is a perennial cause of wars, for example. But few note (in print) how population pressure guarantees competitions for resources, as when Syria’s population growth rate of 2.4% per year in 2010 meant doubling every 30 years. Their economy could not keep up. The corrupt government then used ruthless methods to repress protests, which metastasized into hyper-complex civil wars that eventually displaced fully half of Syria’s pre-war population. At least one-fourth fled the country entirely, spreading their chaos into neighboring nations and even far-away Europe. But most people, including policy makers, focus on particular political actors like Bashar al-Assad and Vladimir Putin rather than on long-term causes of such conflicts.
Whoever wins in Syria today will inherit a country mortally wounded by the incredible damage caused by civil wars, which are fueled especially by corruptions of governance and authoritarian law. And who will fund research on corruptions of governance? Not, usually, governments. But many of them will fund projects to study how to corrupt other governments, or how to exploit policy makers therein, or how to damage adversary economies, rather than on searching for solutions to the common problems facing all of us during this difficult century.
Meanwhile, every nation that currently owns nuclear weapons is modernizing (a.k.a. improving the lethal potentials of their arsenals) while condemning any other nation that tries to join that ‘club.’ This, even though leading thinkers (like former Secretaries of State Henry Kissinger and George Schultz, former Secretary of Defense William Perry, Senator Sam Nunn and others) point out that the system we have built is unstable. 16 In short, the doomsday MAD system 17 that we built with the best of intentions will eventually result in a nuclear war if we wait long enough. Escalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war. Does this danger dominate threat analysis today? No. Catching or killing endless retail terrorists consumes far more attention, time and money.
The nuclear arms control regime that brought a bit more stability and a lot more surveillance, verification, and other measures to restrain the nuclear arms race for decades is unraveling now. Most attention, time and resources are devoted instead to endless wars against untold millions of unemployed and often unemployable teenaged males from failed states. These dispossessed and poorly educated men were created by population pressure. Now they are inspired by militant religious enthusiasts to attack corrupt governments there, here, and increasingly everywhere. None of those problems can be solved by more, or more sophisticated nuclear or other WMD.
What applies to nuclear weapons applies with modifications to chemical, biological and other, more novel weapons of mass destruction. For various reasons, the logic of forbidding these whole categories of weapons from “civilized” arsenals resulted in non-perfect, but very effective bans on most national programs in biological and chemical weapons, overseen by international agencies with resources to detect violations and initiate response contingencies. 18 With nuclear WMD, however, we sleep-walk toward an Armageddon no general or politician says they want.
Conclusions
1. Russian efforts to disrupt the 2016 American Presidential election were remarkably successful, consequential and efficient. Rather than trying to excel at this kind of destructive, black politics, we should implement radical defensive measures, and reflect on how many other nations view America today because of our historic efforts to manipulate their elections, particularly in Latin America. Strategic intelligence attempts to discern those situations that call for response in kind (like an armed invasion of home territories) and other situations where violating core American principles like truth and democracy is short sighted and self-destructive.
2. Many other countries are getting into the social media disinformation game, and black money is as important as black propaganda. But China is still the rising threat, not Russia. Black money can also swing elections, and the US Supreme Court’s “Citizens United” decision opened a door for hostile foreign powers to inject black money into our political system using simple cutouts (like the NRA) 19. Retaliating in kind would be an imprudent, expensive and ineffective response. Still, defense against black money is as essential to democracy as defense against black propaganda through novel methods. Defense is not always retaliation. In each case we should recognize that China is the rising threat, not Russia, even though Russia was first to injure us by those particular methods. China’s police-state is way ahead of us in using facial recognition, artificial intelligence, and other technologies to repress dissent at home ever more efficiently. Such tools will be used against free people everywhere if we do not react. 20
3. A Civil Defense style response to Russian election meddling is more appropriate than “active measures” in response. This differs from cyberwarfare, or kinetic warfare, where offensive capability is as important as defensive resilience. Points one and two allude to this distinction. Some kinds of attack, like direct warfare, call for ruthless, active responses, and practitioners know that you cannot win with defense only. However, when evil information operations threaten freedom and democracy, there are other ways of responding that do less damage to core American values and are arguably more effective over time. Examples would be public education, the Voice of America and similar efforts that spread light and truth, rather than darkness and evil through lies and deception. Strategic intelligence highlights such distinctions.
4. A new emphasis on strategic intelligence over tactical urgencies is also required. This should include specific things like:
a. Recognizing that Climate Change is a new ‘common enemy’ of humankind, and that Information Operations are critical to effective countermeasures. Such “operations” include classical, intelligence led, pro-democracy tools like VoA, but also substantial public education at home to counter ordinary ignorance and to stress that the emerging world faces problems that no nation can solve alone. Among other virtues, this would put America on the side of youth worldwide, who are quite tired of elders arguing while their climate burns and nature decays in a thousand ways.
b. Recognizing that Population Pressure is killing the living system of our planet slowly and fueling wars worldwide while we focus on religious and political enemies, and “triggering events.” Those tactical issues do not disappear; most must be dealt with, but they often obscure ultimate causes and therefore prevent effective solutions. As noted earlier, population pressure is one of the four most important ultimate causes of global conflict today, but discussing it is virtually taboo in practical politics. Yet immigration issues that are driven by population pressure are dominating domestic politics in ever more countries, while the “Sixth Great Extinction” endangers everyone everywhere. 21 Strategic intelligence does not eliminate a single terrorist, evil dictator or rogue-state nuclear weapon, but it does highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms of them. In practical medicine this distinction between symptoms and causes is absolutely critical, and it should be in intelligence analysis as well. 22
c. Examples of Syria and Myanmar Other papers have described in detail how such forces contributed to the collapse of Syria (growth rate in 2010 = 2.4% preceded by four years of climate change-driven drought that intersected with a very corrupt and authoritarian regime, and rising militant religion). 23 That collapse spread population pressure to many other nations including much of Europe. Myanmar faces a similar dilemma, where even genocide of southeastern Rohingya brought no lasting peace, while civil wars fester to the north. Tactical intelligence to “win Myanmar’s wars” does nothing to help it solve its long-term problems, which are becoming profound.
I am Vice President of an “International Society for the Comparative Study of Civilizations.” One perennial topic of concern is whether our emerging global civilization is in a “decay phase” where corrupt and authoritarian leaders fiddle while modern “Romes” burn. People who studied the collapse of ancient civilizations intensively report that sometimes their elites had ample warning of problems to come, from very good sources, but simply ignored them because to consider various taboos was too disturbing to their worldviews. 24 , 25 It is our task today to avoid such catastrophes, and courageous strategic intelligence is a way that intelligence professionals can educate policy makers to do the right things, both morally and practically. | There is
zero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i
c
to both
Obama and Trump
Reasonable differences of informed opinion remain about whether these were successful
but they probably were
But
some other meta-goals of the Russian effort
were unambiguously successful, regardless of whether they tipped the election
Doubt in the legitimacy of
basic institutions, like of our i
c
law enforcement
the D
o
J
and other departments was increased
Democratic intelligence systems depend on
evidence-based analysis and
objective presentation to policy makers of “truth”
When these
are shattered
hell can result
Already a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings
Many factors contribute to the “Developing Global Crisis”
include at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law
These are powerful causes of wars over long periods of history
But they typically underlie more obvious proximate causes of each particular conflict
competition for resources is a perennial cause of wars
every nation that currently owns nuc
s is modernizing
the system
is unstable
MAD
built with the best of intentions will eventually result in
nuclear war
Escalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war
What applies to nuc
s applies with modifications to chemical, biological and other
novel w
m
d
With
WMD
we sleep-walk toward
Armageddon
while
climate burns and nature decays
Population Pressure is killing the living system of our planet
and fueling wars worldwide
the “Sixth Great Extinction” endangers everyone everywhere
intel
does not eliminate a single
dictator or rogue-state
but
does highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms
ample warning of problems to come, from
good sources
avoid such catastrophes, and
intel
professionals can educate policy makers to do the right things | 242 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks",
"role": "system"
},
{
"content": "There is now zero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire intelligence community to both Presidents Obama and Trump. 1 Reasonable differences of informed opinion remain about whether these were successful enough to swing the election from Hillary Clinton to Mr. Trump, but they probably were according to Kathleen Hall Jamieson, Director of the Annenberg Public Policy Center at the University of Pennsylvania. 2 Obviously, if Russian efforts were enough to swing an American Presidential election this would be a profound political effect at minimal cost in national security expenditures (for the Russians anyway). But I want to focus for a moment on some other meta-goals of the Russian effort that were unambiguously successful, regardless of whether they tipped the election to Mr. Trump.\na. Doubt in the validity of electoral process in America in general was greatly amplified.\nb. Doubt in the legitimacy of many basic institutions, like of our intelligence community, law enforcement, media, the Department of Justice and other departments was increased.\nc. Polarization of politics in America, already severe, was also increased, and\nd. Hatred among various American communities was increased. Sewing division, uncertainty and illegitimacy were arguably among Russia’s main strategic goals.\ne. Relationships with long-term allies were put under great stress. NATO suffered. This was another Russian goal. Fractured US relations with Canada and Mexico were bonuses.\nf. The term “Fake News” became a meme, with pernicious effects far beyond the obvious. Democratic intelligence systems depend on concepts like evidence-based analysis and on objective presentation to policy makers of “truth” as best that can be discerned. When these fundamental concepts are shattered, trust evaporates, and hell can result.\nSpecial Counsel Robert Mueller’s report on these topics has not been released to the American public at the time of writing this paper, but it will probably have significant effects on both academic conclusions and on practical politics in America. I can still conclude with high confidence now that this was one of the most effective applications in recent history of what we call “psychological operations” and what Russians often call “active measures.” 3\nThe general dilemma of election meddling with practical consequences for neighbors is nothing new to Eastern Europe. They have dealt with Russian (and other) meddling for a very long time. America is just now learning the hard way why so many people in so many countries hate us for past acts when we manipulated their elections. That negative reaction is a vivid example of blowback from an intelligence practitioner’s or policy maker’s point of view. Policy makers chose a somewhat evil method (manipulating a neighbors’ elections) to support some allegedly good end (like containing communism) but found that in the long run, this also creates much hatred of your country. Hatred has national security consequences too. The internet evolution to distributed networks of information flow, and the special power of “social media” to mold public perceptions of reality are technical aspects of this problem, but the moral dilemma is eternal.\nExcellent technical reviews of evolving methods of manipulation through social media and other emerging fora can be found by Matteo Bonfanti of Zurich’s ETS, and James Pamment et al. of Lund University and Sweden’s Civil Contingencies Agency (MSB). 4 They describe new techniques so exotic they require a new language, as noted in this quote from Bonfanti’s article. “They include socio-cognitive and psychographic hacking via dark advertisement, social hacking, band wagoning, digital disinformation and fakes, exploitation via bots, botnets and sock-puppets, trolling and flaming.” 5 Greg Treverton has also written about these challenges for the Swedish Defense University. 6 Finally, the Canadian Security Intelligence Service published a very good workshop overview on such techniques, but as is their practice all papers are anonymous. 7 By tapping into and enflaming local prejudices, the Russians spread lies faster and further than truth* broadcast by establishment, mainstream media. 8\nShould we promote truth and goodness in our public diplomacy and private (even secret) actions, or descend to deception and evil in order to gain advantage in our competitions for power, wealth and basic security? Well in war, this seems simple. If you do not win, you and all you love may die young. But in cold wars or normal competitions, this is not simple. Evil methods tend to yield evil outcomes far more often than not. Unintended consequences abound, and are often negative. A similar dilemma occurs when people consider assassination (or “targeted killings”) as they have throughout the centuries and to this day. A particular and interesting contemporary case lies at the intersection of these phenomena, the much-publicized murder of Saudi journalist Jamal Khashoggi at the Saudi Arabian consulate in Istanbul, Turkey, on October 2, 2018. 9\nAssassinations of Journalists and other Suppressions of “Objective Reality”\n“Fake News” can be created in very large volumes by active propaganda efforts funded by states. This can be enhanced by suppression of alternative “narratives.” One very harsh method for that is the murder of journalists who do not comply with official stories. 10 There are many other methods short of murder, like arresting, intimidating or buying off large numbers of independent or critical journalists. And police-states simply own all significant media and fire any journalist who strays from party lines, as occurs in North Korea, China and other police-states these days.\nMany more countries than Saudi Arabia use such tactics. Some are not police-states. For one sad example, NATO ally Turkey, which exposed the especially gruesome and egregious murder of Jamal Khashoggi, imprisons more journalists today than any other nation on earth (as best we can measure such tragedies) due to President Erdoğan’s campaign against supporters of Fetullah Gulan. 11 , 12 Those include many journalists, professors and whole media outlets and schools in Turkey (many since closed down by the government). Turkey is not yet a police-state, although it trends in that direction in 2019. Venezuela provides another example from another continent, as Nicolás Maduro drives his formerly prosperous country towards failed-state conditions. In most such cases, control of the media is a prelude to rot within, but both Russia and China prove that police-states with controlled media can endure for very long periods.\nTrue police-states have always been in the vanguard of these kinds of efforts to control people’s thoughts by controlling the information available to them. But even undeniable and relatively healthy democracies can be guilty of egregious controls on information, including kidnapping journalists and whistleblowers, imprisonment of critics and even assassinations of people who tell inconvenient truths. The saddest examples of all to me come from Israel since it once aspired to be a leader in law and moral thought. Then they kidnapped Israeli citizen Mordecai Vanunu overseas because he told the London Sunday Times about Israel’s clandestine nuclear weapons program. 13 They then held Vanunu in solitary confinement for years lest he share more with international media. After 18 years and release, they convicted him again, for speaking to foreigners about his plight. Meanwhile, every child in neighbor nations knows they could be burned alive by Israel’s nuclear arsenal. Israel has also assassinated many hundreds, probably thousands of its neighbors in Lebanon, Egypt, Syria and Iran. Those include both Palestinians working for their own freedom (and sometimes fighting for their freedom, of course) and nationals of those neighbor states, for many reasons detailed in the incomparable record of Israeli assassinations called “Rise and Kill First,” a book reviewed and approved by IDF military censors in 2018. 14 This source is unique, because it includes the moral dilemmas and angst faced by top Israeli intelligence officials as they decided to kill many, many hundreds of possible enemies of their state. Despite strenuous efforts to avoid injuring bystanders, innocents and witnesses nearby were also killed more often than anyone wants to admit. Hence the intense secrecy about “targeted killing” operations, even in a state that aspires to moral leadership.\nI will not detail assassinations attempted, conducted or enabled by my own United States of America because it continues to deny such things. The official denials are lies, but they are important to keep practitioners loyal to my country that also once aspired to moral leadership.\nIt appears that no nation is immune to the temptation to use unconventional methods to attack people or institutions that appear to be threats to the state. “Unconventional methods” include here (more or less) precise targeted killings and massive information operations against whole populations that exploit every dimension of the internet including social media in particular. What happens to strategic intelligence in this context of constant battle against endless enemies? The short answer is that strategic intelligence is commonly ignored, so the enemies never end.\nNuclear Weapons, other WMD, the Developing Global Crisis, and Strategic Intelligence\nStrategic intelligence is often ignored simply because most of the system is very busy every day with current intelligence about active armed conflicts and responses to the blizzard of modern propaganda operations. Already a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings, far too long-term, a purview of mostly elder analysts on end of career assignments, and is almost never attended to by active policy makers.\nI will illustrate this by reference to something virtually everyone agrees is a real threat to all of human civilization, nuclear and other WMD, which proliferate as we chat about terrorists. I will also describe very briefly a “Developing Global Crisis” that underlies much of the lethal conflict on earth today. Few policy makers or analysts will dispute that each of those causes of threat to everything under heaven are real, and powerful. But little is done about any of those ultimate causes of armed conflicts, because they are very difficult to deal with, and some are protected by “taboos” (like population pressure). Finally, everyone is so busy fighting the perpetual wars they are trapped in. So the wars continue. Intelligence professionals can describe in detail why they so often ignore ultimate threats to focus on the mosquitos that are biting states today.\nMany factors contribute to the “Developing Global Crisis” so descriptions can be cumbersome, but all include at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law. 15 These are powerful causes of wars over long periods of history. They are especially relevant today as we become ever more global, economically and in many other ways. But they typically underlie more obvious proximate causes of each particular conflict. This is especially true of the first two. Everyone knows that competition for resources is a perennial cause of wars, for example. But few note (in print) how population pressure guarantees competitions for resources, as when Syria’s population growth rate of 2.4% per year in 2010 meant doubling every 30 years. Their economy could not keep up. The corrupt government then used ruthless methods to repress protests, which metastasized into hyper-complex civil wars that eventually displaced fully half of Syria’s pre-war population. At least one-fourth fled the country entirely, spreading their chaos into neighboring nations and even far-away Europe. But most people, including policy makers, focus on particular political actors like Bashar al-Assad and Vladimir Putin rather than on long-term causes of such conflicts.\nWhoever wins in Syria today will inherit a country mortally wounded by the incredible damage caused by civil wars, which are fueled especially by corruptions of governance and authoritarian law. And who will fund research on corruptions of governance? Not, usually, governments. But many of them will fund projects to study how to corrupt other governments, or how to exploit policy makers therein, or how to damage adversary economies, rather than on searching for solutions to the common problems facing all of us during this difficult century.\nMeanwhile, every nation that currently owns nuclear weapons is modernizing (a.k.a. improving the lethal potentials of their arsenals) while condemning any other nation that tries to join that ‘club.’ This, even though leading thinkers (like former Secretaries of State Henry Kissinger and George Schultz, former Secretary of Defense William Perry, Senator Sam Nunn and others) point out that the system we have built is unstable. 16 In short, the doomsday MAD system 17 that we built with the best of intentions will eventually result in a nuclear war if we wait long enough. Escalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war. Does this danger dominate threat analysis today? No. Catching or killing endless retail terrorists consumes far more attention, time and money.\nThe nuclear arms control regime that brought a bit more stability and a lot more surveillance, verification, and other measures to restrain the nuclear arms race for decades is unraveling now. Most attention, time and resources are devoted instead to endless wars against untold millions of unemployed and often unemployable teenaged males from failed states. These dispossessed and poorly educated men were created by population pressure. Now they are inspired by militant religious enthusiasts to attack corrupt governments there, here, and increasingly everywhere. None of those problems can be solved by more, or more sophisticated nuclear or other WMD.\nWhat applies to nuclear weapons applies with modifications to chemical, biological and other, more novel weapons of mass destruction. For various reasons, the logic of forbidding these whole categories of weapons from “civilized” arsenals resulted in non-perfect, but very effective bans on most national programs in biological and chemical weapons, overseen by international agencies with resources to detect violations and initiate response contingencies. 18 With nuclear WMD, however, we sleep-walk toward an Armageddon no general or politician says they want.\nConclusions\n1. Russian efforts to disrupt the 2016 American Presidential election were remarkably successful, consequential and efficient. Rather than trying to excel at this kind of destructive, black politics, we should implement radical defensive measures, and reflect on how many other nations view America today because of our historic efforts to manipulate their elections, particularly in Latin America. Strategic intelligence attempts to discern those situations that call for response in kind (like an armed invasion of home territories) and other situations where violating core American principles like truth and democracy is short sighted and self-destructive.\n2. Many other countries are getting into the social media disinformation game, and black money is as important as black propaganda. But China is still the rising threat, not Russia. Black money can also swing elections, and the US Supreme Court’s “Citizens United” decision opened a door for hostile foreign powers to inject black money into our political system using simple cutouts (like the NRA) 19. Retaliating in kind would be an imprudent, expensive and ineffective response. Still, defense against black money is as essential to democracy as defense against black propaganda through novel methods. Defense is not always retaliation. In each case we should recognize that China is the rising threat, not Russia, even though Russia was first to injure us by those particular methods. China’s police-state is way ahead of us in using facial recognition, artificial intelligence, and other technologies to repress dissent at home ever more efficiently. Such tools will be used against free people everywhere if we do not react. 20\n3. A Civil Defense style response to Russian election meddling is more appropriate than “active measures” in response. This differs from cyberwarfare, or kinetic warfare, where offensive capability is as important as defensive resilience. Points one and two allude to this distinction. Some kinds of attack, like direct warfare, call for ruthless, active responses, and practitioners know that you cannot win with defense only. However, when evil information operations threaten freedom and democracy, there are other ways of responding that do less damage to core American values and are arguably more effective over time. Examples would be public education, the Voice of America and similar efforts that spread light and truth, rather than darkness and evil through lies and deception. Strategic intelligence highlights such distinctions.\n4. A new emphasis on strategic intelligence over tactical urgencies is also required. This should include specific things like:\na. Recognizing that Climate Change is a new ‘common enemy’ of humankind, and that Information Operations are critical to effective countermeasures. Such “operations” include classical, intelligence led, pro-democracy tools like VoA, but also substantial public education at home to counter ordinary ignorance and to stress that the emerging world faces problems that no nation can solve alone. Among other virtues, this would put America on the side of youth worldwide, who are quite tired of elders arguing while their climate burns and nature decays in a thousand ways.\nb. Recognizing that Population Pressure is killing the living system of our planet slowly and fueling wars worldwide while we focus on religious and political enemies, and “triggering events.” Those tactical issues do not disappear; most must be dealt with, but they often obscure ultimate causes and therefore prevent effective solutions. As noted earlier, population pressure is one of the four most important ultimate causes of global conflict today, but discussing it is virtually taboo in practical politics. Yet immigration issues that are driven by population pressure are dominating domestic politics in ever more countries, while the “Sixth Great Extinction” endangers everyone everywhere. 21 Strategic intelligence does not eliminate a single terrorist, evil dictator or rogue-state nuclear weapon, but it does highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms of them. In practical medicine this distinction between symptoms and causes is absolutely critical, and it should be in intelligence analysis as well. 22\nc. Examples of Syria and Myanmar Other papers have described in detail how such forces contributed to the collapse of Syria (growth rate in 2010 = 2.4% preceded by four years of climate change-driven drought that intersected with a very corrupt and authoritarian regime, and rising militant religion). 23 That collapse spread population pressure to many other nations including much of Europe. Myanmar faces a similar dilemma, where even genocide of southeastern Rohingya brought no lasting peace, while civil wars fester to the north. Tactical intelligence to “win Myanmar’s wars” does nothing to help it solve its long-term problems, which are becoming profound.\nI am Vice President of an “International Society for the Comparative Study of Civilizations.” One perennial topic of concern is whether our emerging global civilization is in a “decay phase” where corrupt and authoritarian leaders fiddle while modern “Romes” burn. People who studied the collapse of ancient civilizations intensively report that sometimes their elites had ample warning of problems to come, from very good sources, but simply ignored them because to consider various taboos was too disturbing to their worldviews. 24 , 25 It is our task today to avoid such catastrophes, and courageous strategic intelligence is a way that intelligence professionals can educate policy makers to do the right things, both morally and practically.",
"role": "user"
},
{
"content": "There is\nzero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i\nc\nto both\nObama and Trump\nReasonable differences of informed opinion remain about whether these were successful\nbut they probably were\nBut\nsome other meta-goals of the Russian effort\nwere unambiguously successful, regardless of whether they tipped the election\nDoubt in the legitimacy of\nbasic institutions, like of our i\nc\nlaw enforcement\nthe D\no\nJ\nand other departments was increased\nDemocratic intelligence systems depend on\nevidence-based analysis and\nobjective presentation to policy makers of “truth”\nWhen these\nare shattered\nhell can result\nAlready a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings\nMany factors contribute to the “Developing Global Crisis”\ninclude at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law\nThese are powerful causes of wars over long periods of history\nBut they typically underlie more obvious proximate causes of each particular conflict\ncompetition for resources is a perennial cause of wars\nevery nation that currently owns nuc\ns is modernizing\nthe system\nis unstable\nMAD\nbuilt with the best of intentions will eventually result in\nnuclear war\nEscalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war\nWhat applies to nuc\ns applies with modifications to chemical, biological and other\nnovel w\nm\nd\nWith\nWMD\nwe sleep-walk toward\nArmageddon\nwhile\nclimate burns and nature decays\nPopulation Pressure is killing the living system of our planet\nand fueling wars worldwide\nthe “Sixth Great Extinction” endangers everyone everywhere\nintel\ndoes not eliminate a single\ndictator or rogue-state\nbut\ndoes highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms\nample warning of problems to come, from\ngood sources\navoid such catastrophes, and\nintel\nprofessionals can educate policy makers to do the right things",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks | There is
zero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i
c
to both
Obama and Trump
Reasonable differences of informed opinion remain about whether these were successful
but they probably were
But
some other meta-goals of the Russian effort
were unambiguously successful, regardless of whether they tipped the election
Doubt in the legitimacy of
basic institutions, like of our i
c
law enforcement
the D
o
J
and other departments was increased
Democratic intelligence systems depend on
evidence-based analysis and
objective presentation to policy makers of “truth”
When these
are shattered
hell can result
Already a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings
Many factors contribute to the “Developing Global Crisis”
include at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law
These are powerful causes of wars over long periods of history
But they typically underlie more obvious proximate causes of each particular conflict
competition for resources is a perennial cause of wars
every nation that currently owns nuc
s is modernizing
the system
is unstable
MAD
built with the best of intentions will eventually result in
nuclear war
Escalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war
What applies to nuc
s applies with modifications to chemical, biological and other
novel w
m
d
With
WMD
we sleep-walk toward
Armageddon
while
climate burns and nature decays
Population Pressure is killing the living system of our planet
and fueling wars worldwide
the “Sixth Great Extinction” endangers everyone everywhere
intel
does not eliminate a single
dictator or rogue-state
but
does highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms
ample warning of problems to come, from
good sources
avoid such catastrophes, and
intel
professionals can educate policy makers to do the right things | zero doubt
aggressively involved
hack
entire i
c
successful
probably were
goals
legitimacy of
basic institutions
i
c
law enforcement
D
o
J
other departments
increased
intel
depend
ev
-based analysis
objective presentation to policy makers
“truth”
hell can result
Already
tactical
strategic
irrelevant
Global Crisis
four factors
1) Population Pressure
2) Corruption
3) Militant Religion
4) Authoritarian Law
cause
wars
obvious proximate
competition for resources
perennial cause
owns nuc
s
modernizing
unstable
MAD
nuc
war
Escalation chains
individuals
computers
any trigger
thermonuclear war
nuc
s
chem
bio
w
m
d
sleep-walk toward
Armageddon
climate burns
nature decays
Population Pressure
killing the living system of our planet
wars worldwide
Great Extinction
endangers everyone everywhere
intel
single
causes
symptoms
ample warning
good sources
avoid
catastrophe
intel
professionals
educate policy makers
do the right thing | 243 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks",
"role": "system"
},
{
"content": "There is\nzero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i\nc\nto both\nObama and Trump\nReasonable differences of informed opinion remain about whether these were successful\nbut they probably were\nBut\nsome other meta-goals of the Russian effort\nwere unambiguously successful, regardless of whether they tipped the election\nDoubt in the legitimacy of\nbasic institutions, like of our i\nc\nlaw enforcement\nthe D\no\nJ\nand other departments was increased\nDemocratic intelligence systems depend on\nevidence-based analysis and\nobjective presentation to policy makers of “truth”\nWhen these\nare shattered\nhell can result\nAlready a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings\nMany factors contribute to the “Developing Global Crisis”\ninclude at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law\nThese are powerful causes of wars over long periods of history\nBut they typically underlie more obvious proximate causes of each particular conflict\ncompetition for resources is a perennial cause of wars\nevery nation that currently owns nuc\ns is modernizing\nthe system\nis unstable\nMAD\nbuilt with the best of intentions will eventually result in\nnuclear war\nEscalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war\nWhat applies to nuc\ns applies with modifications to chemical, biological and other\nnovel w\nm\nd\nWith\nWMD\nwe sleep-walk toward\nArmageddon\nwhile\nclimate burns and nature decays\nPopulation Pressure is killing the living system of our planet\nand fueling wars worldwide\nthe “Sixth Great Extinction” endangers everyone everywhere\nintel\ndoes not eliminate a single\ndictator or rogue-state\nbut\ndoes highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms\nample warning of problems to come, from\ngood sources\navoid such catastrophes, and\nintel\nprofessionals can educate policy makers to do the right things",
"role": "user"
},
{
"content": "zero doubt\naggressively involved\nhack\nentire i\nc\nsuccessful\nprobably were\ngoals\nlegitimacy of\nbasic institutions\ni\nc\nlaw enforcement\nD\no\nJ\nother departments\nincreased\nintel\ndepend\nev\n-based analysis\nobjective presentation to policy makers\n“truth”\nhell can result\nAlready\ntactical\nstrategic\nirrelevant\nGlobal Crisis\nfour factors\n1) Population Pressure\n2) Corruption\n3) Militant Religion\n4) Authoritarian Law\ncause\nwars\nobvious proximate\ncompetition for resources\nperennial cause\nowns nuc\ns\nmodernizing\nunstable\nMAD\nnuc\nwar\nEscalation chains\nindividuals\ncomputers\nany trigger\nthermonuclear war\nnuc\ns\nchem\nbio\nw\nm\nd\nsleep-walk toward\nArmageddon\nclimate burns\nnature decays\nPopulation Pressure\nkilling the living system of our planet\nwars worldwide\nGreat Extinction\nendangers everyone everywhere\nintel\nsingle\ncauses\nsymptoms\nample warning\ngood sources\navoid\ncatastrophe\nintel\nprofessionals\neducate policy makers\ndo the right thing",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks
Highlight Ratio: 0.4 | There is
zero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i
c
to both
Obama and Trump
Reasonable differences of informed opinion remain about whether these were successful
but they probably were
But
some other meta-goals of the Russian effort
were unambiguously successful, regardless of whether they tipped the election
Doubt in the legitimacy of
basic institutions, like of our i
c
law enforcement
the D
o
J
and other departments was increased
Democratic intelligence systems depend on
evidence-based analysis and
objective presentation to policy makers of “truth”
When these
are shattered
hell can result
Already a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings
Many factors contribute to the “Developing Global Crisis”
include at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law
These are powerful causes of wars over long periods of history
But they typically underlie more obvious proximate causes of each particular conflict
competition for resources is a perennial cause of wars
every nation that currently owns nuc
s is modernizing
the system
is unstable
MAD
built with the best of intentions will eventually result in
nuclear war
Escalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war
What applies to nuc
s applies with modifications to chemical, biological and other
novel w
m
d
With
WMD
we sleep-walk toward
Armageddon
while
climate burns and nature decays
Population Pressure is killing the living system of our planet
and fueling wars worldwide
the “Sixth Great Extinction” endangers everyone everywhere
intel
does not eliminate a single
dictator or rogue-state
but
does highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms
ample warning of problems to come, from
good sources
avoid such catastrophes, and
intel
professionals can educate policy makers to do the right things | Russia
hack
our
election
goals
were
successful
Doubt in the legitimacy of
our i
c
law enforcement
increased
Democratic intel
depend on
ev
-based analysis and
objective presentation
of “truth”
When
shattered
hell can result
Population Pressure
Corruption
Militant Religion, and
Authoritarian Law
cause
wars over long periods
underlie
proximate causes
MAD
will
result in
nuc
war
Escalation chains
by computers
decide whether any trigger
becomes
thermonuclear war
applies
to chem
bio
and other
w
m
d
we sleep-walk toward
Armageddon
climate burns and nature decays
killing
our planet
and fueling wars worldwide
Extinction” endangers everyone
intel
does not eliminate a single
rogue-state
but
does highlight responses
dealing with
causes
instead of
symptoms
ample warning
from
good sources
avoid
catastrophe
intel
can educate policy makers to do the right thing | 244 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Russia-induced intelligence politicization causes extinction by undermining responses to nuclear, WMD, and environmental risks\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "There is\nzero doubt that Russia was aggressively involved in hacking our 2016 US election, as cited by our entire i\nc\nto both\nObama and Trump\nReasonable differences of informed opinion remain about whether these were successful\nbut they probably were\nBut\nsome other meta-goals of the Russian effort\nwere unambiguously successful, regardless of whether they tipped the election\nDoubt in the legitimacy of\nbasic institutions, like of our i\nc\nlaw enforcement\nthe D\no\nJ\nand other departments was increased\nDemocratic intelligence systems depend on\nevidence-based analysis and\nobjective presentation to policy makers of “truth”\nWhen these\nare shattered\nhell can result\nAlready a dwarf compared to tactical intelligence, strategic intelligence becomes irrelevant to most taskings\nMany factors contribute to the “Developing Global Crisis”\ninclude at least four factors: 1) Population Pressure, 2) Corruptions of Governance, 3) Militant Religion, and 4) Authoritarian Law\nThese are powerful causes of wars over long periods of history\nBut they typically underlie more obvious proximate causes of each particular conflict\ncompetition for resources is a perennial cause of wars\nevery nation that currently owns nuc\ns is modernizing\nthe system\nis unstable\nMAD\nbuilt with the best of intentions will eventually result in\nnuclear war\nEscalation chains, key individuals and decisions by computers will decide whether any triggering event becomes a general thermonuclear war\nWhat applies to nuc\ns applies with modifications to chemical, biological and other\nnovel w\nm\nd\nWith\nWMD\nwe sleep-walk toward\nArmageddon\nwhile\nclimate burns and nature decays\nPopulation Pressure is killing the living system of our planet\nand fueling wars worldwide\nthe “Sixth Great Extinction” endangers everyone everywhere\nintel\ndoes not eliminate a single\ndictator or rogue-state\nbut\ndoes highlight responses that can be effective at dealing with the causes of conflicts instead of just symptoms\nample warning of problems to come, from\ngood sources\navoid such catastrophes, and\nintel\nprofessionals can educate policy makers to do the right things",
"role": "user"
},
{
"content": "Russia\nhack\nour\nelection\ngoals\nwere\nsuccessful\nDoubt in the legitimacy of\nour i\nc\nlaw enforcement\nincreased\nDemocratic intel\ndepend on\nev\n-based analysis and\nobjective presentation\nof “truth”\nWhen\nshattered\nhell can result\nPopulation Pressure\nCorruption\nMilitant Religion, and\nAuthoritarian Law\ncause\nwars over long periods\nunderlie\nproximate causes\nMAD\nwill\nresult in\nnuc\nwar\nEscalation chains\nby computers\ndecide whether any trigger\nbecomes\nthermonuclear war\napplies\nto chem\nbio\nand other\nw\nm\nd\nwe sleep-walk toward\nArmageddon\nclimate burns and nature decays\nkilling\nour planet\nand fueling wars worldwide\nExtinction” endangers everyone\nintel\ndoes not eliminate a single\nrogue-state\nbut\ndoes highlight responses\ndealing with\ncauses\ninstead of\nsymptoms\nample warning\nfrom\ngood sources\navoid\ncatastrophe\nintel\ncan educate policy makers to do the right thing",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Goes nuclear | Since February 2014, the Russian leadership has been in a de facto war mode with regard to the United States. The Kremlin saw the developments in Ukraine that led to the ouster of former president Viktor Yanukovych as a threefold threat: a U.S.-supported political invasion of Russia’s vital strategic buffer, an attempt to prevent Moscow-led integration in post-Soviet Eurasia, and a move to build a barrier between Russia and the rest of Europe. Russian President Vladimir Putin, taken by surprise, responded with the use of force in Ukraine—first to secure Crimea for Russia and then to protect a rebel stronghold in Donbass. The events that followed have developed into a virtual Russo-American war—but a different kind of war compared to those the countries have fought in the past.
The crisis over Ukraine put an end to a quarter century of cooperative relations between Russia and the West and resulted in Russia’s confrontation with the United States and its estrangement from Europe. This confrontation has often been labeled a second Cold War.1 The analogy, however, is flawed: the world has changed too much since the 1980s to suggest that today’s antagonism is merely a revival of an old conflict. The new confrontation is better described as a Hybrid War—a term which, like its predecessor, is capitalized here to highlight its distinct place in the history of international relations. This time, the U.S.-Russia conflict is not central to the world system, but, nevertheless, its outcome will help shape the future of that system.
The current Hybrid War is a conflict essentially between Russia and the United States over the issue of the world order. It is not the result of misunderstanding or miscalculation but rather the opposite; Russia, in particular, has a deliberate outcome in mind. Moscow is pursuing a set of objectives—the most important of which is to reassert its role as a great power with a global reach. In Europe, specifically, it seeks to prevent NATO from moving forward into former Soviet territory, particularly Ukraine. As for Ukraine itself, the Kremlin wants it to serve as a buffer between Russia and NATO. Russia has important objectives outside of Europe as well, including in the Middle East. Since September 2015, Moscow has been waging a military campaign in Syria. The main purpose of the intervention—apart from the immediate need to prevent a major victory for Islamist extremists—was to return Russia to the regional and global stage as an active geopolitical player with considerable military capabilities. Russian actions in these and other areas therefore undermine the United States’ global dominance of the post–Cold War period, even though the Russian Federation (unlike the Soviet Union) does not seek to impose its own model on the world.
Even as Russia opposes U.S. global hegemony and favors a more distributed balance of power among several major nations (including itself), the United States feels the challenge to the international liberal order that it began building after the end of World War II and has dominated since the end of the Cold War. As long as all major powers, including China and Russia, subscribed to the rules and norms of that order—and, in China’s case, also benefited from it2—it was a genuine Pax Americana: a state of peace among the major powers, who all deferred to the United States. With Russia’s breakout from the post–Cold War system, that unique period of peaceful relations among the principal players is now history.
Even though the scale of the current conflict is much smaller, the stakes are high once more. For the Kremlin, this is a battle for survival—of Russia’s status as an independent player capable of defining and defending its interests and of the Russian leadership, which has been personally targeted by Western financial sanctions and various public accusations ranging from corruption to war crimes. Originally, Moscow believed that this conflict would be a short-term problem, but it now appears to be more prolonged than previously anticipated and may take a generation to resolve.
FEATURES OF THE HYBRID WAR
This Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment. International interaction is no longer restricted by walls or other state-imposed barriers. Traditional distinctions between strategy and tactics have been all but erased. The hybrid warriors include many more players than was the case during the Cold War—from national governments and transnational corporations to nongovernmental actors and even private individuals.
The war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies. Just as important to the Hybrid War is economics, which has been the key driver of globalization that paralleled the rise of these innovative information technologies. The prominence of the U.S. media and the United States’ immense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries.3
Politically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions (for example, restrictions on travel, seizure of assets, imprisonment, or deportation) on opponents. The existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before. Cross-border promotion of democracy and support for the color revolutions that dominated the 2000s (for example, the 2003 Rose Revolution in Georgia and the 2004 Orange Revolution in Ukraine) have now found counterparts in emerging solidarity among those who espouse more conservative and traditionalist values, such as political systems based on authoritarian models and strict national sovereignty.4
Military power is not out of the picture—though its use is different than in the Cold War. The static standoff of million-strong armies in Europe and the long shadow of the nuclear arms race have drawn down or faded. Nuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries, but, to date, this standoff bears no resemblance in either scale or scope to the forces that faced each other during the Cold War. The main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace. Even more than in the past, however, the onus is on national leaderships to minimize the number of casualties, ideally to zero.
Russian military strategists had developed the concept of hybrid warfare even before the actual conflict broke out in earnest between the United States and Russia in early 2014. Analyzing the experience of the post-Soviet color revolutions and the 2011 Arab Spring, Chief of the General Staff Valery Gerasimov wrote in February 2013 that the “consequences of new conflicts are comparable to those of a real war”; in many cases, nonmilitary methods “are substantially more effective than the power of arms,” and greater emphasis is placed on “political, economic, information, humanitarian, and other nonmilitary means” and “covert military measures,” including “information warfare and actions by special forces.” In this environment, “overt use of military force, often in the form of peacekeeping or crisis management, takes place only at a certain stage, mainly to achieve final success in a conflict.” With regard to the U.S.-Russia confrontation, another key feature has surfaced: asymmetry between the sides’ capabilities.
POWER ASYMMETRIES AND ASYMMETRIC ACTIONS
Although Gerasimov was referring to a hybrid war when discussing new means and methods of warfare, this analysis uses the newly fashionable term to describe the current U.S.-Russia confrontation. Unlike its Cold War predecessor, this conflict is asymmetrical. At least since the 1970s, the Soviet Union was the United States’ equal in terms of both nuclear and conventional military power. Even beyond its own vast land mass and immediate sphere of influence in Eastern Europe, it wielded considerable ideological power in many Western countries and in the Third World and presided over a system of alliances in Africa, Asia, Europe, Latin America, and the Middle East. The Russian Federation, by contrast, has few formal allies, no satellite states, and a handful of protectorates, if one includes the self-proclaimed states of Abkhazia, Donbass, South Ossetia, and Transnistria. It has no ideology to compare with the comprehensive dogma of Marxism-Leninism, and although it is still a nuclear superpower, it lags far behind the United States in non-nuclear military capabilities. Economically, Russia—with its estimated 1.5 percent of the global gross domestic product—is a dwarf.
Neither the balance nor the correlation of forces, however, will determine the outcome of this confrontation. Despite the glaring asymmetries in the national power of the two sides of the conflict, the course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the United States or both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War.
The United States is going through a triple crisis of its political system, exemplified but not caused by the arrival of President Donald Trump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system.
Russia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future after the eventual departure of its figurehead. Putin’s Kremlin is already working on a political transition that would rejuvenate the elite and improve its competence and performance, but, at the same time, Russian society is also changing and Putin’s heirs cannot take its support for granted. Gross inequality, sluggish economic growth, low vertical mobility, and high-level corruption will present a range of serious challenges to the future Russian leadership.
The eventual outcome of the Hybrid War could be reminiscent of the downfall of the Soviet Union, which was far less the result of the U.S.-Soviet Cold War than of a misguided effort to reform the Soviet Union itself. Russia might break down and break up again, or it might decide on a foreign policy more geared toward its economic needs than to a certain concept of world order. As for the United States, it might decide to limit its global commitments and redesign its international role as the world’s preeminent but no longer dominant state. Yet, in doing so, it will need to accept that its change in status will come with a certain price and that it will not be able to take advantage of the benefits of the position it once enjoyed.
Asymmetries in power lead to asymmetric actions, which as Gerasimov suggested are intended to “neutralize the enemy’s superiority in warfare” or “identify and exploit the enemy’s vulnerabilities.”5 By an order of magnitude—or more—Russia is outgunned, outmanned, and outspent by the combined forces of the United States and its allies. To stay in the fight, it must rely on its few comparative advantages and seek to use them to maximum effect. These advantages include the geographical proximity of some of the main theaters of operation, such as Crimea and eastern Ukraine, where Russia has escalation dominance; the Russian political system, which allows for secretive, swift, and decisive action; and Moscow’s willingness to take much higher risks in view of the disproportionally higher stakes involved for the Russian leadership and a national culture that historically has tolerated higher losses in defense or protection of the Motherland. Through swift decisions and actions, made without prior warning, Russia is capable of surprising its adversaries and keeping them off-balance. This situation promises an uncertain, hard-to-predict, and risky environment, where miscalculation can lead to incidents or collisions that, in turn, lead to escalation. Granted, these incidents would be of a different kind than the tank standoff at Berlin’s Checkpoint Charlie in late October 1961 or the Cuban Missile Crisis barely a year later. Escalation resulting from miscalculation would not be automatic, but the wider damage it could cause needs to be taken seriously.
AVOIDING MISTAKES LEADING TO ESCALATION
The Hybrid War is highly dynamic and, so far, has no agreed-upon rules. In this sense, it resembles the Cold War of the early 1950s rather than that of the 1970s. However, it is possible, up to a point, to avoid military escalation during the Hybrid War. U.S.-Russian antagonism does not mean that the two countries’ interests are in total opposition. Unlike in the second half of the twentieth century, neither party envisions a real shooting war against its adversary and neither wants to allow the situation to become uncontrollable. The most obvious ways to manage the confrontation are incident prevention, confidence building, and arms control.
Incident prevention, on the face of it, should be easy. Since the early 1970s, Moscow and Washington have had agreements in place to avoid incidents, which in the Cold War days carried the risk of escalation to nuclear levels. Effective prevention requires a degree of professionalism, adequate safety measures, and reliable channels of communications. However, during a Hybrid War, these preconditions cannot be taken for granted. Acting from a position of relative weakness, Russia is likely to compensate for its inferior overall strength by raising the stakes of confrontation. | Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U
S
This Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment
Traditional distinctions between strategy and tactics have been all but erased
The war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies
The prominence of the U.S. media and the U
S
immense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries
Politically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions
on opponents
The existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before
Nuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries
The main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace
in Africa, Asia, Europe, Latin America, and the Middle East
Neither the balance nor the correlation of forces, however, will determine the outcome of this confrontation
the course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U
S
or both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War
The U
S
is going through a triple crisis of its political system, exemplified but not caused by
Trump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system
Russia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future
Russian society is also changing and Putin’s heirs cannot take its support for granted
Russia
must rely on
escalation dominance
swift, and decisive action; and
willingness to take
high
risks
without
warning
This
promises an uncertain
and risky environment, where miscalc
can lead to incidents or collisions that, in turn, lead to escalation
not
automatic
neither party envisions a real
war
wants to allow the situation to become uncontrollable
to manage
confrontation
Incident prevention
should be easy
requires
professionalism
However, during a Hybrid War, these
cannot be taken for granted
raising the stakes of confrontation | 245 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Goes nuclear",
"role": "system"
},
{
"content": "Since February 2014, the Russian leadership has been in a de facto war mode with regard to the United States. The Kremlin saw the developments in Ukraine that led to the ouster of former president Viktor Yanukovych as a threefold threat: a U.S.-supported political invasion of Russia’s vital strategic buffer, an attempt to prevent Moscow-led integration in post-Soviet Eurasia, and a move to build a barrier between Russia and the rest of Europe. Russian President Vladimir Putin, taken by surprise, responded with the use of force in Ukraine—first to secure Crimea for Russia and then to protect a rebel stronghold in Donbass. The events that followed have developed into a virtual Russo-American war—but a different kind of war compared to those the countries have fought in the past. \nThe crisis over Ukraine put an end to a quarter century of cooperative relations between Russia and the West and resulted in Russia’s confrontation with the United States and its estrangement from Europe. This confrontation has often been labeled a second Cold War.1 The analogy, however, is flawed: the world has changed too much since the 1980s to suggest that today’s antagonism is merely a revival of an old conflict. The new confrontation is better described as a Hybrid War—a term which, like its predecessor, is capitalized here to highlight its distinct place in the history of international relations. This time, the U.S.-Russia conflict is not central to the world system, but, nevertheless, its outcome will help shape the future of that system. \nThe current Hybrid War is a conflict essentially between Russia and the United States over the issue of the world order. It is not the result of misunderstanding or miscalculation but rather the opposite; Russia, in particular, has a deliberate outcome in mind. Moscow is pursuing a set of objectives—the most important of which is to reassert its role as a great power with a global reach. In Europe, specifically, it seeks to prevent NATO from moving forward into former Soviet territory, particularly Ukraine. As for Ukraine itself, the Kremlin wants it to serve as a buffer between Russia and NATO. Russia has important objectives outside of Europe as well, including in the Middle East. Since September 2015, Moscow has been waging a military campaign in Syria. The main purpose of the intervention—apart from the immediate need to prevent a major victory for Islamist extremists—was to return Russia to the regional and global stage as an active geopolitical player with considerable military capabilities. Russian actions in these and other areas therefore undermine the United States’ global dominance of the post–Cold War period, even though the Russian Federation (unlike the Soviet Union) does not seek to impose its own model on the world. \nEven as Russia opposes U.S. global hegemony and favors a more distributed balance of power among several major nations (including itself), the United States feels the challenge to the international liberal order that it began building after the end of World War II and has dominated since the end of the Cold War. As long as all major powers, including China and Russia, subscribed to the rules and norms of that order—and, in China’s case, also benefited from it2—it was a genuine Pax Americana: a state of peace among the major powers, who all deferred to the United States. With Russia’s breakout from the post–Cold War system, that unique period of peaceful relations among the principal players is now history. \nEven though the scale of the current conflict is much smaller, the stakes are high once more. For the Kremlin, this is a battle for survival—of Russia’s status as an independent player capable of defining and defending its interests and of the Russian leadership, which has been personally targeted by Western financial sanctions and various public accusations ranging from corruption to war crimes. Originally, Moscow believed that this conflict would be a short-term problem, but it now appears to be more prolonged than previously anticipated and may take a generation to resolve. \nFEATURES OF THE HYBRID WAR \nThis Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment. International interaction is no longer restricted by walls or other state-imposed barriers. Traditional distinctions between strategy and tactics have been all but erased. The hybrid warriors include many more players than was the case during the Cold War—from national governments and transnational corporations to nongovernmental actors and even private individuals. \nThe war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies. Just as important to the Hybrid War is economics, which has been the key driver of globalization that paralleled the rise of these innovative information technologies. The prominence of the U.S. media and the United States’ immense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries.3 \nPolitically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions (for example, restrictions on travel, seizure of assets, imprisonment, or deportation) on opponents. The existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before. Cross-border promotion of democracy and support for the color revolutions that dominated the 2000s (for example, the 2003 Rose Revolution in Georgia and the 2004 Orange Revolution in Ukraine) have now found counterparts in emerging solidarity among those who espouse more conservative and traditionalist values, such as political systems based on authoritarian models and strict national sovereignty.4 \nMilitary power is not out of the picture—though its use is different than in the Cold War. The static standoff of million-strong armies in Europe and the long shadow of the nuclear arms race have drawn down or faded. Nuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries, but, to date, this standoff bears no resemblance in either scale or scope to the forces that faced each other during the Cold War. The main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace. Even more than in the past, however, the onus is on national leaderships to minimize the number of casualties, ideally to zero. \nRussian military strategists had developed the concept of hybrid warfare even before the actual conflict broke out in earnest between the United States and Russia in early 2014. Analyzing the experience of the post-Soviet color revolutions and the 2011 Arab Spring, Chief of the General Staff Valery Gerasimov wrote in February 2013 that the “consequences of new conflicts are comparable to those of a real war”; in many cases, nonmilitary methods “are substantially more effective than the power of arms,” and greater emphasis is placed on “political, economic, information, humanitarian, and other nonmilitary means” and “covert military measures,” including “information warfare and actions by special forces.” In this environment, “overt use of military force, often in the form of peacekeeping or crisis management, takes place only at a certain stage, mainly to achieve final success in a conflict.” With regard to the U.S.-Russia confrontation, another key feature has surfaced: asymmetry between the sides’ capabilities. \nPOWER ASYMMETRIES AND ASYMMETRIC ACTIONS \nAlthough Gerasimov was referring to a hybrid war when discussing new means and methods of warfare, this analysis uses the newly fashionable term to describe the current U.S.-Russia confrontation. Unlike its Cold War predecessor, this conflict is asymmetrical. At least since the 1970s, the Soviet Union was the United States’ equal in terms of both nuclear and conventional military power. Even beyond its own vast land mass and immediate sphere of influence in Eastern Europe, it wielded considerable ideological power in many Western countries and in the Third World and presided over a system of alliances in Africa, Asia, Europe, Latin America, and the Middle East. The Russian Federation, by contrast, has few formal allies, no satellite states, and a handful of protectorates, if one includes the self-proclaimed states of Abkhazia, Donbass, South Ossetia, and Transnistria. It has no ideology to compare with the comprehensive dogma of Marxism-Leninism, and although it is still a nuclear superpower, it lags far behind the United States in non-nuclear military capabilities. Economically, Russia—with its estimated 1.5 percent of the global gross domestic product—is a dwarf. \nNeither the balance nor the correlation of forces, however, will determine the outcome of this confrontation. Despite the glaring asymmetries in the national power of the two sides of the conflict, the course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the United States or both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War. \nThe United States is going through a triple crisis of its political system, exemplified but not caused by the arrival of President Donald Trump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system. \nRussia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future after the eventual departure of its figurehead. Putin’s Kremlin is already working on a political transition that would rejuvenate the elite and improve its competence and performance, but, at the same time, Russian society is also changing and Putin’s heirs cannot take its support for granted. Gross inequality, sluggish economic growth, low vertical mobility, and high-level corruption will present a range of serious challenges to the future Russian leadership. \nThe eventual outcome of the Hybrid War could be reminiscent of the downfall of the Soviet Union, which was far less the result of the U.S.-Soviet Cold War than of a misguided effort to reform the Soviet Union itself. Russia might break down and break up again, or it might decide on a foreign policy more geared toward its economic needs than to a certain concept of world order. As for the United States, it might decide to limit its global commitments and redesign its international role as the world’s preeminent but no longer dominant state. Yet, in doing so, it will need to accept that its change in status will come with a certain price and that it will not be able to take advantage of the benefits of the position it once enjoyed. \nAsymmetries in power lead to asymmetric actions, which as Gerasimov suggested are intended to “neutralize the enemy’s superiority in warfare” or “identify and exploit the enemy’s vulnerabilities.”5 By an order of magnitude—or more—Russia is outgunned, outmanned, and outspent by the combined forces of the United States and its allies. To stay in the fight, it must rely on its few comparative advantages and seek to use them to maximum effect. These advantages include the geographical proximity of some of the main theaters of operation, such as Crimea and eastern Ukraine, where Russia has escalation dominance; the Russian political system, which allows for secretive, swift, and decisive action; and Moscow’s willingness to take much higher risks in view of the disproportionally higher stakes involved for the Russian leadership and a national culture that historically has tolerated higher losses in defense or protection of the Motherland. Through swift decisions and actions, made without prior warning, Russia is capable of surprising its adversaries and keeping them off-balance. This situation promises an uncertain, hard-to-predict, and risky environment, where miscalculation can lead to incidents or collisions that, in turn, lead to escalation. Granted, these incidents would be of a different kind than the tank standoff at Berlin’s Checkpoint Charlie in late October 1961 or the Cuban Missile Crisis barely a year later. Escalation resulting from miscalculation would not be automatic, but the wider damage it could cause needs to be taken seriously. \nAVOIDING MISTAKES LEADING TO ESCALATION \nThe Hybrid War is highly dynamic and, so far, has no agreed-upon rules. In this sense, it resembles the Cold War of the early 1950s rather than that of the 1970s. However, it is possible, up to a point, to avoid military escalation during the Hybrid War. U.S.-Russian antagonism does not mean that the two countries’ interests are in total opposition. Unlike in the second half of the twentieth century, neither party envisions a real shooting war against its adversary and neither wants to allow the situation to become uncontrollable. The most obvious ways to manage the confrontation are incident prevention, confidence building, and arms control. \nIncident prevention, on the face of it, should be easy. Since the early 1970s, Moscow and Washington have had agreements in place to avoid incidents, which in the Cold War days carried the risk of escalation to nuclear levels. Effective prevention requires a degree of professionalism, adequate safety measures, and reliable channels of communications. However, during a Hybrid War, these preconditions cannot be taken for granted. Acting from a position of relative weakness, Russia is likely to compensate for its inferior overall strength by raising the stakes of confrontation.",
"role": "user"
},
{
"content": "Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U\nS\nThis Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment\nTraditional distinctions between strategy and tactics have been all but erased\nThe war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies\nThe prominence of the U.S. media and the U\nS\nimmense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries\nPolitically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions\non opponents\nThe existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before\nNuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries\nThe main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace\nin Africa, Asia, Europe, Latin America, and the Middle East\nNeither the balance nor the correlation of forces, however, will determine the outcome of this confrontation\nthe course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U\nS\nor both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War\nThe U\nS\nis going through a triple crisis of its political system, exemplified but not caused by\nTrump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system\nRussia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future\nRussian society is also changing and Putin’s heirs cannot take its support for granted\nRussia\nmust rely on\nescalation dominance\nswift, and decisive action; and\nwillingness to take\nhigh\nrisks\nwithout\nwarning\nThis\npromises an uncertain\nand risky environment, where miscalc\ncan lead to incidents or collisions that, in turn, lead to escalation\nnot\nautomatic\nneither party envisions a real\nwar\nwants to allow the situation to become uncontrollable\nto manage\nconfrontation\nIncident prevention\nshould be easy\nrequires\nprofessionalism\nHowever, during a Hybrid War, these\ncannot be taken for granted\nraising the stakes of confrontation",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Goes nuclear | Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U
S
This Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment
Traditional distinctions between strategy and tactics have been all but erased
The war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies
The prominence of the U.S. media and the U
S
immense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries
Politically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions
on opponents
The existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before
Nuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries
The main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace
in Africa, Asia, Europe, Latin America, and the Middle East
Neither the balance nor the correlation of forces, however, will determine the outcome of this confrontation
the course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U
S
or both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War
The U
S
is going through a triple crisis of its political system, exemplified but not caused by
Trump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system
Russia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future
Russian society is also changing and Putin’s heirs cannot take its support for granted
Russia
must rely on
escalation dominance
swift, and decisive action; and
willingness to take
high
risks
without
warning
This
promises an uncertain
and risky environment, where miscalc
can lead to incidents or collisions that, in turn, lead to escalation
not
automatic
neither party envisions a real
war
wants to allow the situation to become uncontrollable
to manage
confrontation
Incident prevention
should be easy
requires
professionalism
However, during a Hybrid War, these
cannot be taken for granted
raising the stakes of confrontation | de facto war mode
Hybrid War
global
borderless environment
prime importance
cyber
a
i
propaganda
fake news
info
tech
weapons of choice
Hybrid War
information
discredit
sanction
stimulation
political change
elections
common info
space
political war
easier
Nuclear
risks
miscalc
conventional forces
tech
blur
war
and peace
Africa
Asia
Europe
Latin America
Mid
East
outcome
nonlinear
asymmetrical
domestic developments
U
S
crisis
political system
Trump
foreign policy
escalation dominance
decisive action
high
risks
warning
uncertain
and risky environment
miscalc
collisions
escalat
automatic
envisions
wants
uncontrollable
should
professionalism
Hybrid War
cannot
taken for granted
raising the stakes of confrontation | 246 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Goes nuclear",
"role": "system"
},
{
"content": "Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U\nS\nThis Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment\nTraditional distinctions between strategy and tactics have been all but erased\nThe war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies\nThe prominence of the U.S. media and the U\nS\nimmense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries\nPolitically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions\non opponents\nThe existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before\nNuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries\nThe main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace\nin Africa, Asia, Europe, Latin America, and the Middle East\nNeither the balance nor the correlation of forces, however, will determine the outcome of this confrontation\nthe course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U\nS\nor both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War\nThe U\nS\nis going through a triple crisis of its political system, exemplified but not caused by\nTrump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system\nRussia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future\nRussian society is also changing and Putin’s heirs cannot take its support for granted\nRussia\nmust rely on\nescalation dominance\nswift, and decisive action; and\nwillingness to take\nhigh\nrisks\nwithout\nwarning\nThis\npromises an uncertain\nand risky environment, where miscalc\ncan lead to incidents or collisions that, in turn, lead to escalation\nnot\nautomatic\nneither party envisions a real\nwar\nwants to allow the situation to become uncontrollable\nto manage\nconfrontation\nIncident prevention\nshould be easy\nrequires\nprofessionalism\nHowever, during a Hybrid War, these\ncannot be taken for granted\nraising the stakes of confrontation",
"role": "user"
},
{
"content": "de facto war mode\nHybrid War\nglobal\nborderless environment\nprime importance\ncyber\na\ni\npropaganda\nfake news\ninfo\ntech\nweapons of choice\nHybrid War\ninformation\ndiscredit\nsanction\nstimulation\npolitical change\nelections\ncommon info\nspace\npolitical war\neasier\nNuclear\nrisks\nmiscalc\nconventional forces\ntech\nblur\nwar\nand peace\nAfrica\nAsia\nEurope\nLatin America\nMid\nEast\noutcome\nnonlinear\nasymmetrical\ndomestic developments\nU\nS\ncrisis\npolitical system\nTrump\nforeign policy\nescalation dominance\ndecisive action\nhigh\nrisks\nwarning\nuncertain\nand risky environment\nmiscalc\ncollisions\nescalat\nautomatic\nenvisions\nwants\nuncontrollable\nshould\nprofessionalism\nHybrid War\ncannot\ntaken for granted\nraising the stakes of confrontation",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Goes nuclear
Highlight Ratio: 0.1 | Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U
S
This Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment
Traditional distinctions between strategy and tactics have been all but erased
The war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies
The prominence of the U.S. media and the U
S
immense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries
Politically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions
on opponents
The existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before
Nuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries
The main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace
in Africa, Asia, Europe, Latin America, and the Middle East
Neither the balance nor the correlation of forces, however, will determine the outcome of this confrontation
the course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U
S
or both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War
The U
S
is going through a triple crisis of its political system, exemplified but not caused by
Trump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system
Russia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future
Russian society is also changing and Putin’s heirs cannot take its support for granted
Russia
must rely on
escalation dominance
swift, and decisive action; and
willingness to take
high
risks
without
warning
This
promises an uncertain
and risky environment, where miscalc
can lead to incidents or collisions that, in turn, lead to escalation
not
automatic
neither party envisions a real
war
wants to allow the situation to become uncontrollable
to manage
confrontation
Incident prevention
should be easy
requires
professionalism
However, during a Hybrid War, these
cannot be taken for granted
raising the stakes of confrontation | Hybrid War
is
fought in a
global
borderless environment
From cyber
and
a
i
to
propaganda and fake news
the weapons of choice
are
information
includes
interference in elections
Nuclear
risks of miscalc
derive from
conventional forces. A
standoff has reemerged
new
tech
blur
distinction between war
and peace
in Africa, Asia, Europe, Latin America, and the Mid
East
a nonlinear
asymmetrical conflict
will result
Russia
must rely on
escalation dominance
decisive action; and
high
risks
without
warning
an uncertain
risky environment, where miscalc
lead to
collisions that
escalat
prevention
should be easy
However, during
Hybrid War
cannot be taken for granted | 247 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Goes nuclear\n\nHighlight Ratio: 0.1",
"role": "system"
},
{
"content": "Since February 2014, the Russian leadership has been in a de facto war mode with regard to the U\nS\nThis Hybrid War’s most distinguishing feature is that it is being fought in a truly global, virtually borderless environment\nTraditional distinctions between strategy and tactics have been all but erased\nThe war is being fought simultaneously in a number of spheres, on different levels, and in the never-ending, twenty-four-hour news cycle. This aspect of warfare is particularly true of the field of information, which is of prime importance in the Information Age that emerged with the end of the Cold War. From cyber conflicts and the use of artificial intelligence to the predominance of propaganda and fake news, the main battles of the Hybrid War are taking place outside of the purely physical realm and in the domain of new information technologies\nThe prominence of the U.S. media and the U\nS\nimmense financial power give it a huge advantage in both fields. As a result, the weapons of choice in the Hybrid War are those that use information and economic power to discredit and sanction one’s adversaries\nPolitically, the Hybrid War includes the outside stimulation of political changes in other countries through street activism and the promotion of specific values, parties, or popular movements. It has been characterized by interference in elections, political transitions, and other political processes, including various efforts to hack sensitive information, spread compromising or damaging materials and fake news, encourage character assassinations, and impose personal and other noneconomic sanctions\non opponents\nThe existence of a common information space makes waging political warfare on foreign territory much easier and more attractive than ever before\nNuclear deterrence between Russia and the West remains in place but at lower and more stable levels than during the Cold War. Today’s risks of miscalculation derive from potential incidents involving conventional forces. A token military standoff has reemerged along Russia’s border with NATO countries\nThe main focus is on developing new military technologies and novel means and ways of prosecuting warfare—from outer space to cyberspace—that blur or eliminate the distinction between wartime and peacetime. Like its predecessor, the Hybrid War is a war in the time of peace\nin Africa, Asia, Europe, Latin America, and the Middle East\nNeither the balance nor the correlation of forces, however, will determine the outcome of this confrontation\nthe course of events is not predetermined. As a nonlinear, highly asymmetrical conflict, the outcome likely will result from domestic developments in Russia or the U\nS\nor both. Both countries are facing serious problems that could prove decisive in the final calculations of the Hybrid War\nThe U\nS\nis going through a triple crisis of its political system, exemplified but not caused by\nTrump and the virulent domestic opposition to him and his policies. A crisis of social values lies beneath this political crisis and points to a widening gap between the more liberal and the largely conservative parts of the country. At the same time, the United States faces a crisis within its own foreign policy as it struggles to reconcile the conflict between the more inward-looking U.S. national interest and the international liberal order of the U.S.-led global system\nRussia, though outwardly stable, is approaching its own major crisis as the political regime created by Putin faces an uncertain future\nRussian society is also changing and Putin’s heirs cannot take its support for granted\nRussia\nmust rely on\nescalation dominance\nswift, and decisive action; and\nwillingness to take\nhigh\nrisks\nwithout\nwarning\nThis\npromises an uncertain\nand risky environment, where miscalc\ncan lead to incidents or collisions that, in turn, lead to escalation\nnot\nautomatic\nneither party envisions a real\nwar\nwants to allow the situation to become uncontrollable\nto manage\nconfrontation\nIncident prevention\nshould be easy\nrequires\nprofessionalism\nHowever, during a Hybrid War, these\ncannot be taken for granted\nraising the stakes of confrontation",
"role": "user"
},
{
"content": "Hybrid War\nis\nfought in a\nglobal\nborderless environment\nFrom cyber\nand\na\ni\nto\npropaganda and fake news\nthe weapons of choice\nare\ninformation\nincludes\ninterference in elections\nNuclear\nrisks of miscalc\nderive from\nconventional forces. A\nstandoff has reemerged\nnew\ntech\nblur\ndistinction between war\nand peace\nin Africa, Asia, Europe, Latin America, and the Mid\nEast\na nonlinear\nasymmetrical conflict\nwill result\nRussia\nmust rely on\nescalation dominance\ndecisive action; and\nhigh\nrisks\nwithout\nwarning\nan uncertain\nrisky environment, where miscalc\nlead to\ncollisions that\nescalat\nprevention\nshould be easy\nHowever, during\nHybrid War\ncannot be taken for granted",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing. | FISA gives the government an extraordinarily powerful surveillance tool. Under FISA, the government can ask a secret court, the Foreign Intelligence Surveillance Court, for approval to wiretap individuals to obtain “foreign intelligence information.” As compared to ordinary criminal wiretaps, for which the government must establish probable cause that the wiretap will yield evidence of a particularcrime, FISA surveillance is based on relaxed standards, allowing the government to conduct surveillance with fewer restraints. In recent years, the government has relied on FISA to deploy an array of novel and intrusive surveillance techniques — implicating the privacy rights of countless Americans who have never been suspected of any crime.
Disturbingly, the government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing their Fourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud — in a prosecution that was ultimately dismissed. And in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s about as far from espionage or national security as you can get.
To make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable.
Normally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations by government agents. But when it comes to FISA surveillance, defendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.
The Department of Justice Inspector General’s recent report on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.
Indeed, in 2000, the government confessed that 75 surveillance applications it had submitted to the FISA court contained misstatements and omissions of material facts. In 2014, a press report revealed that five prominent Muslim Americans, including one who had served in the Bush administration, were apparently targeted for surveillance.
The one-sided and secretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses. As our amicus brief in Mr. Osseily’s case explains, courts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.
While the courts must do more to uphold the rights of defendants subject to FISA surveillance, Congress also has a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance. | FISA gives the government an extraordinarily powerful surveillance tool
the government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing
Fourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud
And in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s
as far from espionage or national security as you can get.
To make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable
Normally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations
But when it comes to FISA
defendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.
The
I
G
report on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.
in 2000, the government confessed that 75
applications it had submitted to the FISA court contained misstatements and omissions
The one-sided
secretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses
courts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.
Congress
has a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance | 248 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing.",
"role": "system"
},
{
"content": "FISA gives the government an extraordinarily powerful surveillance tool. Under FISA, the government can ask a secret court, the Foreign Intelligence Surveillance Court, for approval to wiretap individuals to obtain “foreign intelligence information.” As compared to ordinary criminal wiretaps, for which the government must establish probable cause that the wiretap will yield evidence of a particularcrime, FISA surveillance is based on relaxed standards, allowing the government to conduct surveillance with fewer restraints. In recent years, the government has relied on FISA to deploy an array of novel and intrusive surveillance techniques — implicating the privacy rights of countless Americans who have never been suspected of any crime.\nDisturbingly, the government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing their Fourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud — in a prosecution that was ultimately dismissed. And in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s about as far from espionage or national security as you can get.\nTo make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable.\nNormally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations by government agents. But when it comes to FISA surveillance, defendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.\nThe Department of Justice Inspector General’s recent report on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.\nIndeed, in 2000, the government confessed that 75 surveillance applications it had submitted to the FISA court contained misstatements and omissions of material facts. In 2014, a press report revealed that five prominent Muslim Americans, including one who had served in the Bush administration, were apparently targeted for surveillance. \nThe one-sided and secretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses. As our amicus brief in Mr. Osseily’s case explains, courts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.\nWhile the courts must do more to uphold the rights of defendants subject to FISA surveillance, Congress also has a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance.",
"role": "user"
},
{
"content": "FISA gives the government an extraordinarily powerful surveillance tool\nthe government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing\nFourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud\nAnd in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s\nas far from espionage or national security as you can get.\nTo make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable\nNormally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations\nBut when it comes to FISA\ndefendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.\nThe\nI\nG\nreport on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.\nin 2000, the government confessed that 75\napplications it had submitted to the FISA court contained misstatements and omissions\nThe one-sided\nsecretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses\ncourts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.\nCongress\nhas a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing. | FISA gives the government an extraordinarily powerful surveillance tool
the government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing
Fourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud
And in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s
as far from espionage or national security as you can get.
To make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable
Normally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations
But when it comes to FISA
defendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.
The
I
G
report on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.
in 2000, the government confessed that 75
applications it had submitted to the FISA court contained misstatements and omissions
The one-sided
secretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses
courts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.
Congress
has a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance | extraordinarily powerful surveillance tool
using
run-of-the-mill criminal investigations against Americans
Fourth Amendment rights
mortgage fraud
wire fraud
food-stamp fraud
as far from espionage or national security as you can get.
immense hurdles
vindicating
ev
review the government’s application
commonsense requirement
bedrock due process protection
assistance of counsel
best positioned
misrepresentations
FISA
never get that opportunity
even when
40 years since FISA was enacted
no defendant has ever been able to review the government’s FISA application.
I
G
report
exactly why that’s a problem
shocking number of misstatements and omissions
adviser to the now-president
other
s less connected
75
applications
misstatements
omissions
one-sided
secretive nature
lack of disclosure to criminal defendants, encourages these abuses
must give defendants access to FISA materials consistent with due process
Congress
has a responsibility
when
relies on FISA in a criminal prosecution
have a right to review the applications and orders that led to the surveillance | 249 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing.",
"role": "system"
},
{
"content": "FISA gives the government an extraordinarily powerful surveillance tool\nthe government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing\nFourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud\nAnd in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s\nas far from espionage or national security as you can get.\nTo make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable\nNormally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations\nBut when it comes to FISA\ndefendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.\nThe\nI\nG\nreport on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.\nin 2000, the government confessed that 75\napplications it had submitted to the FISA court contained misstatements and omissions\nThe one-sided\nsecretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses\ncourts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.\nCongress\nhas a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance",
"role": "user"
},
{
"content": "extraordinarily powerful surveillance tool\nusing\nrun-of-the-mill criminal investigations against Americans\nFourth Amendment rights\nmortgage fraud\nwire fraud\nfood-stamp fraud\nas far from espionage or national security as you can get.\nimmense hurdles\nvindicating\nev\nreview the government’s application\ncommonsense requirement\nbedrock due process protection\nassistance of counsel\nbest positioned\nmisrepresentations\nFISA\nnever get that opportunity\neven when\n40 years since FISA was enacted\nno defendant has ever been able to review the government’s FISA application.\nI\nG\nreport\nexactly why that’s a problem\nshocking number of misstatements and omissions\nadviser to the now-president\nother\ns less connected\n75\napplications\nmisstatements\nomissions\none-sided\nsecretive nature\nlack of disclosure to criminal defendants, encourages these abuses\nmust give defendants access to FISA materials consistent with due process\nCongress\nhas a responsibility\nwhen\nrelies on FISA in a criminal prosecution\nhave a right to review the applications and orders that led to the surveillance",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing.
Highlight Ratio: 0.2 | FISA gives the government an extraordinarily powerful surveillance tool
the government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing
Fourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud
And in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s
as far from espionage or national security as you can get.
To make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable
Normally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations
But when it comes to FISA
defendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.
The
I
G
report on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.
in 2000, the government confessed that 75
applications it had submitted to the FISA court contained misstatements and omissions
The one-sided
secretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses
courts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.
Congress
has a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance | when the government uses
ev
in a criminal case, defendants have the right to review the
application
a commonsense requirement and
bedrock due process
The defendant
counsel, is best positioned to identify
misrepresentations
But
FISA
defendants never get that
The
I
G
exposed
shocking
misstatements
If this
can happen in
an adviser to the
president, one can
imagine
abuses that affect other
s less connected
lack of disclosure
encourages these abuses
Congress
has
responsibility to
make clear that, when
the government relies on FISA in a criminal prosecution, defendants have a right to review
applications | 250 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: The plan ensures defendants’ counsel can challenge FISA warrants underlying criminal evidence. This creates a backstop against inappropriate policing.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "FISA gives the government an extraordinarily powerful surveillance tool\nthe government is increasingly using these broad and intrusive spying powers in run-of-the-mill criminal investigations against Americans, circumventing\nFourth Amendment rights. For instance, the government recently relied on FISA surveillance to charge a man in Washington D.C. with crimes related to mortgage fraud. It used FISA to spy on Temple University physics professor, Xiaoxing Xi, before charging him with wire fraud\nAnd in Michigan, the government revealed that it used FISA surveillance in a prosecution for food-stamp fraud. That’s\nas far from espionage or national security as you can get.\nTo make matters worse, people caught up in FISA prosecutions face immense hurdles in vindicating their due process rights and holding the government accountable\nNormally, when the government uses wiretap evidence in a criminal case, defendants have the right to review the government’s application in order to challenge any errors and omissions. This is a commonsense requirement and a bedrock due process protection: The defendant, with the assistance of counsel, is best positioned to identify potential misrepresentations\nBut when it comes to FISA\ndefendants never get that opportunity — even when the government relies on FISA surveillance in ordinary criminal prosecutions. In the 40 years since FISA was enacted, no defendant has ever been able to review the government’s FISA application.\nThe\nI\nG\nreport on FISA abuses shows exactly why that’s a problem. The report exposed a shocking number of misstatements and omissions in the government’s FISA applications to surveil former Trump campaign advisor Carter Page. If this abuse can happen in the case of an adviser to the now-president, one can only imagine the surveillance abuses that affect other individuals less connected than Page.\nin 2000, the government confessed that 75\napplications it had submitted to the FISA court contained misstatements and omissions\nThe one-sided\nsecretive nature of the FISA process, combined with the lack of disclosure to criminal defendants, encourages these abuses\ncourts must give defendants access to FISA materials consistent with due process, as Congress anticipated when it enacted FISA.\nCongress\nhas a responsibility to reform the statute to make clear that, whenever the government relies on FISA in a criminal prosecution, defendants have a right to review the applications and orders that led to the surveillance",
"role": "user"
},
{
"content": "when the government uses\nev\nin a criminal case, defendants have the right to review the\napplication\na commonsense requirement and\nbedrock due process\nThe defendant\ncounsel, is best positioned to identify\nmisrepresentations\nBut\nFISA\ndefendants never get that\nThe\nI\nG\nexposed\nshocking\nmisstatements\nIf this\ncan happen in\nan adviser to the\npresident, one can\nimagine\nabuses that affect other\ns less connected\nlack of disclosure\nencourages these abuses\nCongress\nhas\nresponsibility to\nmake clear that, when\nthe government relies on FISA in a criminal prosecution, defendants have a right to review\napplications",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed. | Robert Chesney 22:16
What is it you would want to see? Okay, so we should articulate what you and I know is the countervailing consideration that that acts as the counterweight in this process. Obviously, we're not going to move, not just as a predictive matter, but I'm suggesting as a merits matter, we not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What is it you think would be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance? Are you interested in having, creating more ex ante anticipation of review by creating stronger post hoc tools? Do you want to instead see more ex ante or contemporaneous tools?
Steve Vladeck 25:29
I think this actually dovetails with stuff I've written about more generally, I think national security law is a field of law where ex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application, because you know, who's gonna know who the target is, besides the government? No, I mean, I think the the right way to do this, Bobby, is a series of after the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end. And so instead of a one off Horwitz audit, right, what about regular audits of individual, of a randomly selected cross section of Title I applications?
Robert Chesney 26:26
So for example, taking exactly what's done here, and having this be a statutorily required recurring process that ensures that an agent who's in a huge rush, is under tremendous pressure, puts their life on the line on a regular basis, and is worried about other people's lives, nonetheless knows that, it's not just an easy out where you can blow off something, yeah, Woods Procedures, fill in the blank, I said a few words, now onto the thing I want to work on, but rather knows that there's a decent chance that happens all the time that somebody's case from last year gets pulled for audit and then you get blasted and it looks bad on review. And if you know that's the case then you tick the needle over a little bit into how much time and effort the agents will put into actually following the procedures, which on their face are good procedures, it's the execution we're being told here is shoddy. I think that actually to me, sounds pretty reasonable. You know, I am not moving closer to your view. I think both of us are just going to keep the same relative distance between us in terms of general level of trust. I'm assuming your general level of trust is going down, mine's moving some, we may just be moving in lockstep - a lowering tide lowering all boats proportionally.
Steve Vladeck 27:46
Although I started much lower.
Robert Chesney 27:47
Yeah, you're starting much lower. I'm starting much higher. And but I nonetheless think what you're describing sounds pretty reasonable and frankly, not particularly disruptive. I appreciate that you're focusing on ex post procedures as opposed to what the usual argument you hear people say is, "oh, we gotta put an amicus in there in all settings to litigate the way that America does litigation." All right, hold on, I have some requests for you to go fill out! I've got a lot of paperwork for you now to consider.
Steve Vladeck 28:15
But so I said I had three ideas, and I wanted to sort of get them all on the table.
Robert Chesney 28:18
Okay, that was just one.
Steve Vladeck 28:19
Right. So one is one is mandate, by statute, periodic random audits of Title I applications. The second, and this is gonna sort of be in some tension with what you just said, but in the cases in which there already is authority to appoint an amicus, I would go back to the original proposal in the what was called the Leahy Bill in the 2015 FISA reform, the 2014-2015 FISA reform, and make the amicus not an amicus, make it mandatory. Not in all cases, Bobby, for the reasons you say, but stop having this sort of mechanism where it depends upon the court saying "I need help," right, as opposed to, mandateing the participation, at least in those cases where the statute already allows participation, which is to say, cases raising novel questions of law, or important questions of law under the statute. And then the...
Robert Chesney 29:14
Before the third one, I just want to ask a question about that one.
Steve Vladeck 29:15
Yeah.
Robert Chesney 29:16
Would you would you agree the - both Horowitz reports, very, very problematic findings. They're not about that. That, to me is about the sorts of issues that were raised by...
Steve Vladeck 29:30
That's right. And it's not clear to me that an amicus participating in the FISA court process, even in a Title I case, like a Carter Page type case, would have anywhere near enough information to actually, you know, point out the things that Horowitz can point out from behind the scenes as the Inspector General.
Robert Chesney 29:50
I'd be very opposed to that sort of participation.
Steve Vladeck 29:53
Right. Which is why it should be, you know, sort of, in the cases where the statue already exists, just make it a stronger power. But then third, and this is the one that I think comes back to Title I. So, you know, but I think listeners may not sort of follow this jot and tiddle, that FISA itself provides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute, I think it's 1806(f), 50 USC Section 1806(f), provides at least what looks like it's [it is] supposed to be what's called a Franks Hearing, and this is named after the 1978 Supreme Court case, Franks v. Delaware. And the idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application in order to help the judge determine whether it was invalid. This provision has been read quite narrowly by the courts that have carefully considered it. I think, Bobby, you might remember the seventh circuit's 2014 decision in the Daoud case, which was the most prominent one, where Judge Posner really went after a district judge who had, the district judge basically said in order to know which materials are relevant to the Franks Hearing, I need the security-cleared defense lawyer to be able to see the whole application. And Posner says, that's totally crazy. And read 1806(f) incredibly narrowly. I think it's worth having a conversation, Bobby, about whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application if and when the government's going to use evidence against him in a criminal trial. Now, I should say, I mean, this is not a surprise to anyone, it's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but Bobby, again, I think the hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be meaningfully scrutinized by an Article III, civilian, you know, non-FISA Court judge in the context of a motion to suppress, you know, I have to think that would have a salutary effect on the rigor and procedural propriety of what the government does.
Robert Chesney 32:40
So would you feel equally comfortable if, so, the status quo is that it's clear, the judge certainly has authority to consider it - it's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file. Would you be more or less satisfied if there was an amicus in that context so that it was not the defendant's own lawyer? Because you put the defendant's own lawyer in a pretty impossible position if they get exposed - first of all, the majority of these cases, they're not going to be in a position to have the clearances, and it's going to gum up the works to try to get them to that position.
Steve Vladeck 33:18
Is that true? Is that true that in a majority of these cases, the defendant doesn't have the security clear defense counsel?
Robert Chesney 33:24
I don't know. I don't know. That's just my assumption. That was...
Steve Vladeck 33:27
My anecdotal experience, Bobby, and anecdotes are not data, but my anecdotal experience, the cases I'm familiar with, all invariably involve security-cleared defense counsel. I'm sorry, I didn't mean to throw you off.
Robert Chesney 33:38
No, that's good. It's a fair point because I was asserting it like it's empirically documented. I don't know. So I'm happy to - let's just assume it's otherwise, they've all got clearance. They also have a duty to zealously represent their clients and you put them in a very difficult, difficult ethical position. They're the ones best situated to look for the problems, of course, because they know the case best, and amicus is a huge disadvantage. On the other hand, putting them in a position to potentially be exposed to information that the client may not be told is, as we learned in, say, the Moussaoui litigation and some other settings, that can be really problematic, too. But I also think you're just much more likely to actually get such an innovation if the 1806(f) motion to get some adversariality involved ran towards getting one of the standing amicus persons involved. Now that said, I that's actually, I think that's already within the courts authority. And it may be something they could do without any type of obligation.
Steve Vladeck 34:38
But you could make it mandatory. You could make it mandatory. You could also sort of tie the existing amucus provisions in the FISA court to 1806(f) in civilian court. I just think there are there are ways, Bobby, that would not dramatically upset the government's internal processes, where you could put some real teeth into some of the few, sort of, back-end checking mechanisms that are out there that I think would go a long way toward incentivizing, sort of, reduction of shenanigans, reduction of sloppiness, and much more care on the government's part. | Obviously
we not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What
would be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?
ex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application
the right way to do this
is
after the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end
FISA
provides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute
50 USC Section 1806(f)
provides at least what looks like it's [it is] supposed to be what's called a Franks Hearing
the idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application
to help the judge determine whether it was invalid. This
has been read quite narrowly by the courts that have carefully considered it
the seventh circuit
read 1806(f) incredibly narrowly
it's worth having a conversation
about whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application
when the government's going to use evidence against him in a criminal trial
it's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but
the hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be
scrutinized by an Article III
non-FISA
judge in the context of a motion to suppress
that would have a salutary effect on the rigor and procedural propriety of what the government does
it's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file
security-cleared defense counsel
have a duty to zealously represent their clients
They're the
best situated to look for the problems
because they know the case best
amicus is a huge disadvantage
there are ways
that would not dramatically upset the government's internal processes, where you could put
real teeth into
the
back-end checking mechanisms that are out there that
would go a long way toward incentivizing
reduction of shenanigans, reduction of sloppiness, and much more care on the government's part | 251 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed.",
"role": "system"
},
{
"content": "Robert Chesney 22:16 \nWhat is it you would want to see? Okay, so we should articulate what you and I know is the countervailing consideration that that acts as the counterweight in this process. Obviously, we're not going to move, not just as a predictive matter, but I'm suggesting as a merits matter, we not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What is it you think would be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance? Are you interested in having, creating more ex ante anticipation of review by creating stronger post hoc tools? Do you want to instead see more ex ante or contemporaneous tools? \nSteve Vladeck 25:29 \nI think this actually dovetails with stuff I've written about more generally, I think national security law is a field of law where ex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application, because you know, who's gonna know who the target is, besides the government? No, I mean, I think the the right way to do this, Bobby, is a series of after the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end. And so instead of a one off Horwitz audit, right, what about regular audits of individual, of a randomly selected cross section of Title I applications?\nRobert Chesney 26:26 \nSo for example, taking exactly what's done here, and having this be a statutorily required recurring process that ensures that an agent who's in a huge rush, is under tremendous pressure, puts their life on the line on a regular basis, and is worried about other people's lives, nonetheless knows that, it's not just an easy out where you can blow off something, yeah, Woods Procedures, fill in the blank, I said a few words, now onto the thing I want to work on, but rather knows that there's a decent chance that happens all the time that somebody's case from last year gets pulled for audit and then you get blasted and it looks bad on review. And if you know that's the case then you tick the needle over a little bit into how much time and effort the agents will put into actually following the procedures, which on their face are good procedures, it's the execution we're being told here is shoddy. I think that actually to me, sounds pretty reasonable. You know, I am not moving closer to your view. I think both of us are just going to keep the same relative distance between us in terms of general level of trust. I'm assuming your general level of trust is going down, mine's moving some, we may just be moving in lockstep - a lowering tide lowering all boats proportionally. \nSteve Vladeck 27:46 \nAlthough I started much lower. \nRobert Chesney 27:47 \nYeah, you're starting much lower. I'm starting much higher. And but I nonetheless think what you're describing sounds pretty reasonable and frankly, not particularly disruptive. I appreciate that you're focusing on ex post procedures as opposed to what the usual argument you hear people say is, \"oh, we gotta put an amicus in there in all settings to litigate the way that America does litigation.\" All right, hold on, I have some requests for you to go fill out! I've got a lot of paperwork for you now to consider.\nSteve Vladeck 28:15 \nBut so I said I had three ideas, and I wanted to sort of get them all on the table. \nRobert Chesney 28:18 \nOkay, that was just one. \nSteve Vladeck 28:19 \nRight. So one is one is mandate, by statute, periodic random audits of Title I applications. The second, and this is gonna sort of be in some tension with what you just said, but in the cases in which there already is authority to appoint an amicus, I would go back to the original proposal in the what was called the Leahy Bill in the 2015 FISA reform, the 2014-2015 FISA reform, and make the amicus not an amicus, make it mandatory. Not in all cases, Bobby, for the reasons you say, but stop having this sort of mechanism where it depends upon the court saying \"I need help,\" right, as opposed to, mandateing the participation, at least in those cases where the statute already allows participation, which is to say, cases raising novel questions of law, or important questions of law under the statute. And then the...\nRobert Chesney 29:14 \nBefore the third one, I just want to ask a question about that one. \nSteve Vladeck 29:15 \nYeah. \nRobert Chesney 29:16 \nWould you would you agree the - both Horowitz reports, very, very problematic findings. They're not about that. That, to me is about the sorts of issues that were raised by... \nSteve Vladeck 29:30 \nThat's right. And it's not clear to me that an amicus participating in the FISA court process, even in a Title I case, like a Carter Page type case, would have anywhere near enough information to actually, you know, point out the things that Horowitz can point out from behind the scenes as the Inspector General.\nRobert Chesney 29:50 \nI'd be very opposed to that sort of participation.\nSteve Vladeck 29:53 \nRight. Which is why it should be, you know, sort of, in the cases where the statue already exists, just make it a stronger power. But then third, and this is the one that I think comes back to Title I. So, you know, but I think listeners may not sort of follow this jot and tiddle, that FISA itself provides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute, I think it's 1806(f), 50 USC Section 1806(f), provides at least what looks like it's [it is] supposed to be what's called a Franks Hearing, and this is named after the 1978 Supreme Court case, Franks v. Delaware. And the idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application in order to help the judge determine whether it was invalid. This provision has been read quite narrowly by the courts that have carefully considered it. I think, Bobby, you might remember the seventh circuit's 2014 decision in the Daoud case, which was the most prominent one, where Judge Posner really went after a district judge who had, the district judge basically said in order to know which materials are relevant to the Franks Hearing, I need the security-cleared defense lawyer to be able to see the whole application. And Posner says, that's totally crazy. And read 1806(f) incredibly narrowly. I think it's worth having a conversation, Bobby, about whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application if and when the government's going to use evidence against him in a criminal trial. Now, I should say, I mean, this is not a surprise to anyone, it's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but Bobby, again, I think the hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be meaningfully scrutinized by an Article III, civilian, you know, non-FISA Court judge in the context of a motion to suppress, you know, I have to think that would have a salutary effect on the rigor and procedural propriety of what the government does.\nRobert Chesney 32:40 \nSo would you feel equally comfortable if, so, the status quo is that it's clear, the judge certainly has authority to consider it - it's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file. Would you be more or less satisfied if there was an amicus in that context so that it was not the defendant's own lawyer? Because you put the defendant's own lawyer in a pretty impossible position if they get exposed - first of all, the majority of these cases, they're not going to be in a position to have the clearances, and it's going to gum up the works to try to get them to that position. \nSteve Vladeck 33:18 \nIs that true? Is that true that in a majority of these cases, the defendant doesn't have the security clear defense counsel? \nRobert Chesney 33:24 \nI don't know. I don't know. That's just my assumption. That was...\nSteve Vladeck 33:27 \nMy anecdotal experience, Bobby, and anecdotes are not data, but my anecdotal experience, the cases I'm familiar with, all invariably involve security-cleared defense counsel. I'm sorry, I didn't mean to throw you off.\nRobert Chesney 33:38 \nNo, that's good. It's a fair point because I was asserting it like it's empirically documented. I don't know. So I'm happy to - let's just assume it's otherwise, they've all got clearance. They also have a duty to zealously represent their clients and you put them in a very difficult, difficult ethical position. They're the ones best situated to look for the problems, of course, because they know the case best, and amicus is a huge disadvantage. On the other hand, putting them in a position to potentially be exposed to information that the client may not be told is, as we learned in, say, the Moussaoui litigation and some other settings, that can be really problematic, too. But I also think you're just much more likely to actually get such an innovation if the 1806(f) motion to get some adversariality involved ran towards getting one of the standing amicus persons involved. Now that said, I that's actually, I think that's already within the courts authority. And it may be something they could do without any type of obligation. \nSteve Vladeck 34:38 \nBut you could make it mandatory. You could make it mandatory. You could also sort of tie the existing amucus provisions in the FISA court to 1806(f) in civilian court. I just think there are there are ways, Bobby, that would not dramatically upset the government's internal processes, where you could put some real teeth into some of the few, sort of, back-end checking mechanisms that are out there that I think would go a long way toward incentivizing, sort of, reduction of shenanigans, reduction of sloppiness, and much more care on the government's part.",
"role": "user"
},
{
"content": "Obviously\nwe not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What\nwould be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?\nex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application\nthe right way to do this\nis\nafter the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end\nFISA\nprovides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute\n50 USC Section 1806(f)\nprovides at least what looks like it's [it is] supposed to be what's called a Franks Hearing\nthe idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application\nto help the judge determine whether it was invalid. This\nhas been read quite narrowly by the courts that have carefully considered it\nthe seventh circuit\nread 1806(f) incredibly narrowly\nit's worth having a conversation\nabout whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application\nwhen the government's going to use evidence against him in a criminal trial\nit's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but\nthe hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be\nscrutinized by an Article III\nnon-FISA\njudge in the context of a motion to suppress\nthat would have a salutary effect on the rigor and procedural propriety of what the government does\nit's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file\nsecurity-cleared defense counsel\nhave a duty to zealously represent their clients\nThey're the\nbest situated to look for the problems\nbecause they know the case best\namicus is a huge disadvantage\nthere are ways\nthat would not dramatically upset the government's internal processes, where you could put\nreal teeth into\nthe\nback-end checking mechanisms that are out there that\nwould go a long way toward incentivizing\nreduction of shenanigans, reduction of sloppiness, and much more care on the government's part",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed. | Obviously
we not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What
would be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?
ex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application
the right way to do this
is
after the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end
FISA
provides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute
50 USC Section 1806(f)
provides at least what looks like it's [it is] supposed to be what's called a Franks Hearing
the idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application
to help the judge determine whether it was invalid. This
has been read quite narrowly by the courts that have carefully considered it
the seventh circuit
read 1806(f) incredibly narrowly
it's worth having a conversation
about whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application
when the government's going to use evidence against him in a criminal trial
it's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but
the hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be
scrutinized by an Article III
non-FISA
judge in the context of a motion to suppress
that would have a salutary effect on the rigor and procedural propriety of what the government does
it's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file
security-cleared defense counsel
have a duty to zealously represent their clients
They're the
best situated to look for the problems
because they know the case best
amicus is a huge disadvantage
there are ways
that would not dramatically upset the government's internal processes, where you could put
real teeth into
the
back-end checking mechanisms that are out there that
would go a long way toward incentivizing
reduction of shenanigans, reduction of sloppiness, and much more care on the government's part | are not
should not
proposed target
involved in any way
tantamount to just not doing it at all
would
reasonable way
speed and efficiency
without tipping off in advance?
ex ante
much hard
to defend
lawyer
right way
after the fact checking
auditing
hydraulic pressure
cleaner on the front end
procedure
criminal case
defendant
collaterally attack the underlying FISA warrant
50 USC Section 1806(f)
Franks Hearing
look at the underlying application
invalid
quite narrowly by the courts
seventh circuit
incredibly narrowly
actually serving its purpose
pressure on the government
teeth into the
right to
attack the application
when the government's going to use evidence against him in a criminal trial
tiny fraction
hydraulic pressure
scrutinized by a
motion to suppress
salutary effect on the rigor and procedural propriety of what the government does
ex parte
adversariality
duty
zealously represent their clients
best situated to look for the problems
know the case best
amicus
huge disadvantage
not dramatically upset the government's internal processes
real teeth
back-end checking mechanisms
incentivizing
reduction of shenanigans
sloppiness
care | 252 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed.",
"role": "system"
},
{
"content": "Obviously\nwe not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What\nwould be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?\nex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application\nthe right way to do this\nis\nafter the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end\nFISA\nprovides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute\n50 USC Section 1806(f)\nprovides at least what looks like it's [it is] supposed to be what's called a Franks Hearing\nthe idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application\nto help the judge determine whether it was invalid. This\nhas been read quite narrowly by the courts that have carefully considered it\nthe seventh circuit\nread 1806(f) incredibly narrowly\nit's worth having a conversation\nabout whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application\nwhen the government's going to use evidence against him in a criminal trial\nit's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but\nthe hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be\nscrutinized by an Article III\nnon-FISA\njudge in the context of a motion to suppress\nthat would have a salutary effect on the rigor and procedural propriety of what the government does\nit's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file\nsecurity-cleared defense counsel\nhave a duty to zealously represent their clients\nThey're the\nbest situated to look for the problems\nbecause they know the case best\namicus is a huge disadvantage\nthere are ways\nthat would not dramatically upset the government's internal processes, where you could put\nreal teeth into\nthe\nback-end checking mechanisms that are out there that\nwould go a long way toward incentivizing\nreduction of shenanigans, reduction of sloppiness, and much more care on the government's part",
"role": "user"
},
{
"content": "are not\nshould not\nproposed target\ninvolved in any way\ntantamount to just not doing it at all\nwould\nreasonable way\nspeed and efficiency\nwithout tipping off in advance?\nex ante\nmuch hard\nto defend\nlawyer\nright way\nafter the fact checking\nauditing\nhydraulic pressure\ncleaner on the front end\nprocedure\ncriminal case\ndefendant\ncollaterally attack the underlying FISA warrant\n50 USC Section 1806(f)\nFranks Hearing\nlook at the underlying application\ninvalid\nquite narrowly by the courts\nseventh circuit\nincredibly narrowly\nactually serving its purpose\npressure on the government\nteeth into the\nright to\nattack the application\nwhen the government's going to use evidence against him in a criminal trial\ntiny fraction\nhydraulic pressure\nscrutinized by a\nmotion to suppress\nsalutary effect on the rigor and procedural propriety of what the government does\nex parte\nadversariality\nduty\nzealously represent their clients\nbest situated to look for the problems\nknow the case best\namicus\nhuge disadvantage\nnot dramatically upset the government's internal processes\nreal teeth\nback-end checking mechanisms\nincentivizing\nreduction of shenanigans\nsloppiness\ncare",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed.
Highlight Ratio: 0.3 | Obviously
we not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What
would be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?
ex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application
the right way to do this
is
after the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end
FISA
provides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute
50 USC Section 1806(f)
provides at least what looks like it's [it is] supposed to be what's called a Franks Hearing
the idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application
to help the judge determine whether it was invalid. This
has been read quite narrowly by the courts that have carefully considered it
the seventh circuit
read 1806(f) incredibly narrowly
it's worth having a conversation
about whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application
when the government's going to use evidence against him in a criminal trial
it's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but
the hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be
scrutinized by an Article III
non-FISA
judge in the context of a motion to suppress
that would have a salutary effect on the rigor and procedural propriety of what the government does
it's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file
security-cleared defense counsel
have a duty to zealously represent their clients
They're the
best situated to look for the problems
because they know the case best
amicus is a huge disadvantage
there are ways
that would not dramatically upset the government's internal processes, where you could put
real teeth into
the
back-end checking mechanisms that are out there that
would go a long way toward incentivizing
reduction of shenanigans, reduction of sloppiness, and much more care on the government's part | a situation in which the
target
is involved
is
tantamount to
not doing it at all. What
would
enable
speed and efficiency
without tipping off in advance?
ex ante tools are
hard
to defend
the right way
is
after the fact checking
that exert
pressure
to be cleaner on the front end
FISA
provides
what
is
supposed to be
a Franks Hearing
has been read
incredibly narrowly
one of the
ways to
pressure
the government to
have its ducks in a row
is to put
teeth into the
right to
attack the application
when the government's going to use evidence
in a
trial
it's a tiny fraction of FISA
but
the hydraulic pressure of
knowing
this
could
be
scrutinized by a
judge in
a motion to suppress
would have a salutary effect
security-cleared defense counsel
have a duty to
represent their clients
They're
best situated to look for
problems
because they know the case best
amicus is a huge disadvantage
incentivizing
reduction of shenanigans,
of sloppiness, and
more care | 253 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: It threads the needle, causing downstream effects for the entire FISA process without unduly limiting investigation speed.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "Obviously\nwe not only are not going to but should not move to a situation in which the proposed target of the surveillance is involved in any way and I don't know anyone who would argue otherwise because that's [that is] tantamount to just not doing it at all. What\nwould be a reasonable way to enable the requisite speed and efficiency of trying to monitor foreign agents foreign intelligence threats, foreign terrorism threats, without tipping off in advance?\nex ante tools are much harder to defend normatively than ex post tools because, you know, in this context, what would ex ante tools look like? You're not going to put a lawyer in the FISA court to argue against the government's FISA application\nthe right way to do this\nis\nafter the fact checking and auditing mechanisms that exert hydraulic pressure on the process to be cleaner on the front end\nFISA\nprovides a procedure in any criminal case in which the government wants to introduce evidence derived from FISA for the defendant to collaterally attack the underlying FISA warrant. The statute\n50 USC Section 1806(f)\nprovides at least what looks like it's [it is] supposed to be what's called a Franks Hearing\nthe idea is that if the defendant can offer some non-speculative, some sort of plausible reason for thinking that there might be a problem with the warrant application, then he, or at least his security-cleared lawyer, should be allowed to look at the underlying application\nto help the judge determine whether it was invalid. This\nhas been read quite narrowly by the courts that have carefully considered it\nthe seventh circuit\nread 1806(f) incredibly narrowly\nit's worth having a conversation\nabout whether 1806(f) is actually serving its purpose, and whether one of the other ways to put pressure on the government to really have its ducks in a row when it's pursuing a FISA application is to put more teeth into the defendant's right to collaterally attack the application\nwhen the government's going to use evidence against him in a criminal trial\nit's a tiny fraction of FISA cases that ever end up in that context, where the government's trying to bring 1806 evidence, trying to bring FISA evidence into a criminal case triggering 1806(f), but\nthe hydraulic pressure of having that out there, and of the government knowing that this application could, one day, be\nscrutinized by an Article III\nnon-FISA\njudge in the context of a motion to suppress\nthat would have a salutary effect on the rigor and procedural propriety of what the government does\nit's a question of whether the judge does it ex parte which is what always happens, or if instead the judge actually allows some adversariality in the review of the file\nsecurity-cleared defense counsel\nhave a duty to zealously represent their clients\nThey're the\nbest situated to look for the problems\nbecause they know the case best\namicus is a huge disadvantage\nthere are ways\nthat would not dramatically upset the government's internal processes, where you could put\nreal teeth into\nthe\nback-end checking mechanisms that are out there that\nwould go a long way toward incentivizing\nreduction of shenanigans, reduction of sloppiness, and much more care on the government's part",
"role": "user"
},
{
"content": "a situation in which the\ntarget\nis involved\nis\ntantamount to\nnot doing it at all. What\nwould\nenable\nspeed and efficiency\nwithout tipping off in advance?\nex ante tools are\nhard\nto defend\nthe right way\nis\nafter the fact checking\nthat exert\npressure\nto be cleaner on the front end\nFISA\nprovides\nwhat\nis\nsupposed to be\na Franks Hearing\nhas been read\nincredibly narrowly\none of the\nways to\npressure\nthe government to\nhave its ducks in a row\nis to put\nteeth into the\nright to\nattack the application\nwhen the government's going to use evidence\nin a\ntrial\nit's a tiny fraction of FISA\nbut\nthe hydraulic pressure of\nknowing\nthis\ncould\nbe\nscrutinized by a\njudge in\na motion to suppress\nwould have a salutary effect\nsecurity-cleared defense counsel\nhave a duty to\nrepresent their clients\nThey're\nbest situated to look for\nproblems\nbecause they know the case best\namicus is a huge disadvantage\nincentivizing\nreduction of shenanigans,\nof sloppiness, and\nmore care",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes. | Ninety percent of seafood consumed in the United States is imported. Could a U.S. seafood company prohibit the importation of a Thai competitor’s illegally harvested fish? The vast majority of apparel sold in the United States is imported. Could a clothing company exclude from the U.S. market garments that a competitor in Vietnam made with illegal child labor? The answer to these questions is yes: Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against these and other unlawful trade practices.
Although Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language of Section 337(a)(1)(A) — “[u]nfair methods of competition and unfair acts” incident to the importation of products — is profoundly broad, and “unfairness” is not defined in the statute or its legislative history.[1] The ITC has already construed unfairness to include misappropriation of trade secrets, false designation of origin, false labeling, false advertising, violation of the Digital Millennium Copyright Act and antitrust violations.[2] Section 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling products in the United States.
Section 337’s reach is international and its adjudication is fast, making the ITC a compelling venue for a company or interest group opposed to an unlawful practice overseas that affects the United States through trade.[3] Upon institution of an investigation, a complainant can immediately serve the respondents — domestic and foreign — with discovery requests seeking information about their business practices. Respondents failing to appear can be held in default. In this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.
Prevailing on Non-IP Claims Under Section 337
To establish a prima facie case under Section 337(a)(1)(A), a complainant must demonstrate: (1) unfair methods of competition or unfair acts; (2) relating to imported merchandise; and (3) injury to a U.S. industry from such imports.
Unfair Methods or Acts
“What is unfair under Section 337(a)(1)(A)?” is a threshold question for any entity seeking to use the statute to challenge imports that do not infringe statutory IP. Notably, the Senate report accompanying the original passage of the statute explained that “[t]he provision relating to unfair methods of competition in the importation of goods is broad enough to prevent every type and form of unfair practice and is, therefore, a more adequate protection to American industry than any antidumping statute the country has ever had.”[4]
The predecessor court to the U.S. Court of Appeals for the Federal Circuit recognized that Section 337’s prohibitions are:
broad and inclusive and should not be held to be limited to acts coming within the technical definition of unfair methods of competition as applied in some decisions. The importation of articles may involve questions which differ materially from any arising in purely domestic competition, and it is evident from the language used that Congress intended to allow wide discretion in determining what practices are to be regarded as unfair.[5]
The ITC has intervened where alleged unfairness in the importation of articles threatens “the assurance of competitive conditions in the United States economy,”[6] even for the purpose of halting such unfairness in its incipiency.[7] In light of its broad scope as written and interpreted, Section 337(a)(1)(A) is practically unbounded.
Bottom line: If unfair methods or acts overseas threaten the free and fair operation of the U.S. marketplace, Section 337 may apply. For example, it would be unfair for a Malaysian manufacturer exporting to the United States to pay its workers a wage so low that it violates Malaysian law.
Nexus Between Unfairness and Importation
As the ITC has explained:
It is obvious from our traditional role, not to mention our remedial provisions, that Congress intended Section 337 to attack only unfair trade practices which relate to imported products. It then becomes crucial to discern some nexus between unfair methods or acts and importation before this commission has power to act.
A Chinese company manufacturing widgets for the U.S. market would engage in unfairness by defrauding its suppliers into providing it goods and services for free, and such fraud might allow the manufacturer to import the widgets into the United States at an artificially low price. Yet, if the manufacturer pocketed the ill-gotten gains and imported the widgets into the United States at a fair price, the unfairness, while unfortunate for the foreign suppliers, would be too removed from the United States to justify Section 337’s involvement. Likewise, if a U.S. retailer of the Chinese manufacturer’s widgets committed fraud only in the post-importation sale of widgets, Americans might suffer, but the ITC would stand idle, deferring to the jurisdiction of U.S. federal and state courts.
Bottom line: A complainant must connect the alleged unfairness to the importation of articles. For example, show that underpaid workers produced the imported goods, allowing the goods to be sold in the United States at an artificially low price.
Injury to a U.S. Industry
A complainant under Section 337(a)(1)(A) must have a domestic industry that is the target of the unfair methods or acts. “The commission does not adhere to any rigid formula in determining the scope of the domestic industry as it is not precisely defined in the statute, but will examine each case in light of the realities of the marketplace.”[8] Generally speaking, a company with U.S. employees that expends resources in the United States — whether on manufacturing, customer service, research and development, etc. — should be able to establish a domestic industry under Section 337(a)(1)(A).
A complainant must then show substantial injury from the unfairness — either real or threatened. The ITC's injury determinations tend to focus on five factors: (1) the respondent’s volume of imports and penetration into the market; (2) the complainant’s lost sales; (3) underselling by the respondent; (4) the complainant’s declining production, profitability and sales; and (5) harm to goodwill and reputation.[9] The ITC considers three additional criteria in assessing threat of substantial injury: (1) foreign cost advantages and production capacity; (2) the ability of the imported product to undersell the complainant’s product; or (3) substantial foreign manufacturing capacity combined with the respondent’s intention to penetrate the U.S. market.[10] While showing injury (or threat thereof) can entail complex economic analyses, in practice, the ITC has not imposed a high threshold for satisfying this element of Section 337(a)(1)(A).
Bottom line: Connect the unfairness to real or threatened injury. For example, show that the goods imported at an artificially low price have or will likely erode the market share of a company operating in the U.S. market, regardless of whether that company manufactures here.
Hypothetical Cases Under Section 337
Conflict Minerals
Many U.S. consumers would be surprised to learn that their smartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the U.S. Securities and Exchange Commission issued a rule implementing Section 1502 of the Dodd-Frank Act, requiring all publicly traded companies to disclose their use of certain minerals associated with conflict and used in consumer goods.[11] Failure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors who expend resources complying with the law.
Another “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law. In February 2012, the government of the Democratic Republic of the Congo codified a requirement that all mining and mineral trading companies operating in the country perform supply chain due diligence to ensure their purchases are not supporting warring parties in that country. In theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the United States.
Child Labor
The illegality of child labor is widely recognized. The United Nations Convention on the Rights of the Child recites basic human rights for all children, including the right to protection from economic exploitation.[12] Additionally, the Minimum Age Convention requires countries to ensure that children below a certain minimum age (which varies depending on the activity) are not employed.[13] Section 337 may be well-situated for claims based on the importation of articles produced with child labor in contravention of these treaties.
A violation of the national labor laws of the country-of-origin of imported goods could also trigger a Section 337 case. Indeed, “nations that violate [their own] fair labor laws and accepted standards ... [can] accrue an unfairly gained competitive advantage through unfair reduction of the cost of labor, a major input in the cost of production, thereby distorting trade.”[14] This practice has been labeled “social dumping,” defined as the “export of products that owe their competitiveness to low labor standards.”[15] A Section 337 investigation could improve standards for child workers in poor countries by threatening import-driven segments of the U.S. marketplace that traffic in goods made by such children.
Environmental Degradation
The Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife[,] ... any plant,” or “any prohibited wildlife species” that has been “taken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law.”[16] “Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”[17] Collectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.
A Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price. For example, companies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties, [18] or of importing lumber harvested in protected forests.
Food and Drug Safety
Only 2 percent of U.S. seafood imports are inspected by the U.S. Food and Drug Administration before entering the U.S. market.[19] One public interest group found that 33 percent of samples were mislabeled and thus sold to unsuspecting consumers under false, and potentially dangerous, pretenses.[20] Sushi vendors and grocery stores were businesses found most likely to sell mislabeled food. A U.S. seafood company could use Section 337 to challenge the importation of competing, mislabeled fish.
U.S. drug products are increasingly imported from emerging economies, such as India and China, yet the FDA cannot conduct sufficient oversight of foreign plants to ensure compliance with U.S. law. [21] To offer competitive pricing and gain market share at the expense of compliant companies, some overseas plants have abdicated good manufacturing practices and, consequently, caused adulteration of the U.S. drug supply. Drug companies who follow the law could use Section 337 to combat the importation of competing drugs made in substandard facilities, thereby protecting U.S. consumers from unsafe medicines.
Conclusion
Section 337 is a powerful trade remedy administered by a nimble, expert agency. Whether a company or interest group can employ the statute to redress non-IP claims depends on the legal and factual contours of each alleged infraction. If a complainant can establish a prima face case — unfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry — the ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place. | Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against
unlawful trade practices
Although Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language
is profoundly broad
already construed
to include misappropriation of trade secrets, false designation of origin, false labeling
false advertising, violation of the Digital Millennium Copyright Act and antitrust
Section 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling
in the U
S
Section 337’s reach is international and its adjudication is fast, making the ITC a compelling venue
In this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.
Conflict Minerals
the
smartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the
S
E
C
implementing
Dodd-Frank
requiring all publicly traded companies to disclose
use of
minerals associated with conflict and used in consumer goods
Failure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors
Another “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law
In theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U
S
Environmental Degradation
The Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife
any plant
or
prohibited wildlife species
taken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law
“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”
Collectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.
A Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price
companies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties
or of importing lumber harvested in protected forests.
If a complainant can establish a
case
unfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry
the ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place | 254 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes.",
"role": "system"
},
{
"content": "Ninety percent of seafood consumed in the United States is imported. Could a U.S. seafood company prohibit the importation of a Thai competitor’s illegally harvested fish? The vast majority of apparel sold in the United States is imported. Could a clothing company exclude from the U.S. market garments that a competitor in Vietnam made with illegal child labor? The answer to these questions is yes: Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against these and other unlawful trade practices. \nAlthough Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language of Section 337(a)(1)(A) — “[u]nfair methods of competition and unfair acts” incident to the importation of products — is profoundly broad, and “unfairness” is not defined in the statute or its legislative history.[1] The ITC has already construed unfairness to include misappropriation of trade secrets, false designation of origin, false labeling, false advertising, violation of the Digital Millennium Copyright Act and antitrust violations.[2] Section 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling products in the United States. \nSection 337’s reach is international and its adjudication is fast, making the ITC a compelling venue for a company or interest group opposed to an unlawful practice overseas that affects the United States through trade.[3] Upon institution of an investigation, a complainant can immediately serve the respondents — domestic and foreign — with discovery requests seeking information about their business practices. Respondents failing to appear can be held in default. In this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide. \nPrevailing on Non-IP Claims Under Section 337 \nTo establish a prima facie case under Section 337(a)(1)(A), a complainant must demonstrate: (1) unfair methods of competition or unfair acts; (2) relating to imported merchandise; and (3) injury to a U.S. industry from such imports. \nUnfair Methods or Acts \n“What is unfair under Section 337(a)(1)(A)?” is a threshold question for any entity seeking to use the statute to challenge imports that do not infringe statutory IP. Notably, the Senate report accompanying the original passage of the statute explained that “[t]he provision relating to unfair methods of competition in the importation of goods is broad enough to prevent every type and form of unfair practice and is, therefore, a more adequate protection to American industry than any antidumping statute the country has ever had.”[4] \nThe predecessor court to the U.S. Court of Appeals for the Federal Circuit recognized that Section 337’s prohibitions are: \nbroad and inclusive and should not be held to be limited to acts coming within the technical definition of unfair methods of competition as applied in some decisions. The importation of articles may involve questions which differ materially from any arising in purely domestic competition, and it is evident from the language used that Congress intended to allow wide discretion in determining what practices are to be regarded as unfair.[5] \nThe ITC has intervened where alleged unfairness in the importation of articles threatens “the assurance of competitive conditions in the United States economy,”[6] even for the purpose of halting such unfairness in its incipiency.[7] In light of its broad scope as written and interpreted, Section 337(a)(1)(A) is practically unbounded. \nBottom line: If unfair methods or acts overseas threaten the free and fair operation of the U.S. marketplace, Section 337 may apply. For example, it would be unfair for a Malaysian manufacturer exporting to the United States to pay its workers a wage so low that it violates Malaysian law. \nNexus Between Unfairness and Importation \nAs the ITC has explained: \nIt is obvious from our traditional role, not to mention our remedial provisions, that Congress intended Section 337 to attack only unfair trade practices which relate to imported products. It then becomes crucial to discern some nexus between unfair methods or acts and importation before this commission has power to act. \nA Chinese company manufacturing widgets for the U.S. market would engage in unfairness by defrauding its suppliers into providing it goods and services for free, and such fraud might allow the manufacturer to import the widgets into the United States at an artificially low price. Yet, if the manufacturer pocketed the ill-gotten gains and imported the widgets into the United States at a fair price, the unfairness, while unfortunate for the foreign suppliers, would be too removed from the United States to justify Section 337’s involvement. Likewise, if a U.S. retailer of the Chinese manufacturer’s widgets committed fraud only in the post-importation sale of widgets, Americans might suffer, but the ITC would stand idle, deferring to the jurisdiction of U.S. federal and state courts. \nBottom line: A complainant must connect the alleged unfairness to the importation of articles. For example, show that underpaid workers produced the imported goods, allowing the goods to be sold in the United States at an artificially low price. \nInjury to a U.S. Industry \nA complainant under Section 337(a)(1)(A) must have a domestic industry that is the target of the unfair methods or acts. “The commission does not adhere to any rigid formula in determining the scope of the domestic industry as it is not precisely defined in the statute, but will examine each case in light of the realities of the marketplace.”[8] Generally speaking, a company with U.S. employees that expends resources in the United States — whether on manufacturing, customer service, research and development, etc. — should be able to establish a domestic industry under Section 337(a)(1)(A). \nA complainant must then show substantial injury from the unfairness — either real or threatened. The ITC's injury determinations tend to focus on five factors: (1) the respondent’s volume of imports and penetration into the market; (2) the complainant’s lost sales; (3) underselling by the respondent; (4) the complainant’s declining production, profitability and sales; and (5) harm to goodwill and reputation.[9] The ITC considers three additional criteria in assessing threat of substantial injury: (1) foreign cost advantages and production capacity; (2) the ability of the imported product to undersell the complainant’s product; or (3) substantial foreign manufacturing capacity combined with the respondent’s intention to penetrate the U.S. market.[10] While showing injury (or threat thereof) can entail complex economic analyses, in practice, the ITC has not imposed a high threshold for satisfying this element of Section 337(a)(1)(A). \nBottom line: Connect the unfairness to real or threatened injury. For example, show that the goods imported at an artificially low price have or will likely erode the market share of a company operating in the U.S. market, regardless of whether that company manufactures here. \nHypothetical Cases Under Section 337 \nConflict Minerals \nMany U.S. consumers would be surprised to learn that their smartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the U.S. Securities and Exchange Commission issued a rule implementing Section 1502 of the Dodd-Frank Act, requiring all publicly traded companies to disclose their use of certain minerals associated with conflict and used in consumer goods.[11] Failure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors who expend resources complying with the law. \nAnother “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law. In February 2012, the government of the Democratic Republic of the Congo codified a requirement that all mining and mineral trading companies operating in the country perform supply chain due diligence to ensure their purchases are not supporting warring parties in that country. In theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the United States. \nChild Labor \nThe illegality of child labor is widely recognized. The United Nations Convention on the Rights of the Child recites basic human rights for all children, including the right to protection from economic exploitation.[12] Additionally, the Minimum Age Convention requires countries to ensure that children below a certain minimum age (which varies depending on the activity) are not employed.[13] Section 337 may be well-situated for claims based on the importation of articles produced with child labor in contravention of these treaties. \nA violation of the national labor laws of the country-of-origin of imported goods could also trigger a Section 337 case. Indeed, “nations that violate [their own] fair labor laws and accepted standards ... [can] accrue an unfairly gained competitive advantage through unfair reduction of the cost of labor, a major input in the cost of production, thereby distorting trade.”[14] This practice has been labeled “social dumping,” defined as the “export of products that owe their competitiveness to low labor standards.”[15] A Section 337 investigation could improve standards for child workers in poor countries by threatening import-driven segments of the U.S. marketplace that traffic in goods made by such children. \nEnvironmental Degradation \nThe Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife[,] ... any plant,” or “any prohibited wildlife species” that has been “taken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law.”[16] “Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”[17] Collectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States. \nA Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price. For example, companies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties, [18] or of importing lumber harvested in protected forests. \nFood and Drug Safety \nOnly 2 percent of U.S. seafood imports are inspected by the U.S. Food and Drug Administration before entering the U.S. market.[19] One public interest group found that 33 percent of samples were mislabeled and thus sold to unsuspecting consumers under false, and potentially dangerous, pretenses.[20] Sushi vendors and grocery stores were businesses found most likely to sell mislabeled food. A U.S. seafood company could use Section 337 to challenge the importation of competing, mislabeled fish. \nU.S. drug products are increasingly imported from emerging economies, such as India and China, yet the FDA cannot conduct sufficient oversight of foreign plants to ensure compliance with U.S. law. [21] To offer competitive pricing and gain market share at the expense of compliant companies, some overseas plants have abdicated good manufacturing practices and, consequently, caused adulteration of the U.S. drug supply. Drug companies who follow the law could use Section 337 to combat the importation of competing drugs made in substandard facilities, thereby protecting U.S. consumers from unsafe medicines. \nConclusion \nSection 337 is a powerful trade remedy administered by a nimble, expert agency. Whether a company or interest group can employ the statute to redress non-IP claims depends on the legal and factual contours of each alleged infraction. If a complainant can establish a prima face case — unfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry — the ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place.",
"role": "user"
},
{
"content": "Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against\nunlawful trade practices\nAlthough Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language\nis profoundly broad\nalready construed\nto include misappropriation of trade secrets, false designation of origin, false labeling\nfalse advertising, violation of the Digital Millennium Copyright Act and antitrust\nSection 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling\nin the U\nS\nSection 337’s reach is international and its adjudication is fast, making the ITC a compelling venue\nIn this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.\nConflict Minerals\nthe\nsmartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the\nS\nE\nC\nimplementing\nDodd-Frank\nrequiring all publicly traded companies to disclose\nuse of\nminerals associated with conflict and used in consumer goods\nFailure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors\nAnother “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law\nIn theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U\nS\nEnvironmental Degradation\nThe Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife\nany plant\nor\nprohibited wildlife species\ntaken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law\n“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”\nCollectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.\nA Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price\ncompanies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties\nor of importing lumber harvested in protected forests.\nIf a complainant can establish a\ncase\nunfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry\nthe ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes. | Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against
unlawful trade practices
Although Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language
is profoundly broad
already construed
to include misappropriation of trade secrets, false designation of origin, false labeling
false advertising, violation of the Digital Millennium Copyright Act and antitrust
Section 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling
in the U
S
Section 337’s reach is international and its adjudication is fast, making the ITC a compelling venue
In this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.
Conflict Minerals
the
smartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the
S
E
C
implementing
Dodd-Frank
requiring all publicly traded companies to disclose
use of
minerals associated with conflict and used in consumer goods
Failure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors
Another “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law
In theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U
S
Environmental Degradation
The Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife
any plant
or
prohibited wildlife species
taken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law
“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”
Collectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.
A Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price
companies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties
or of importing lumber harvested in protected forests.
If a complainant can establish a
case
unfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry
the ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place | I
T
C
effective weapon
unlawful trade practices
range of unfair acts
profoundly broad
already
trade secrets
designation of origin
labeling
ad
s
Copyright
antitrust
exclusion order
marketing
distributing
selling
U
S
fast
compelling venue
appeal of the U.S. marketplace
fairer
markets
Conflict Minerals
militias
S
E
C
Dodd-Frank
requir
disclose
use of
minerals associated with conflict
consumer goods
comply
337 action
foreign sovereign’s law
reach companies trading in certain minerals mined in Congo
Congolese compliance
U
S
Environmental Degradation
Lacey Act
demonstrating unfairness
unlawful
unfair
fish
wildlife
plant
prohibited wildlife species
foreign law
Section 337 investigation
violation of the Lacey Act
flouted U.S. or foreign environmental law
seafood
lumber
excellent forum
more equitable place | 255 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes.",
"role": "system"
},
{
"content": "Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against\nunlawful trade practices\nAlthough Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language\nis profoundly broad\nalready construed\nto include misappropriation of trade secrets, false designation of origin, false labeling\nfalse advertising, violation of the Digital Millennium Copyright Act and antitrust\nSection 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling\nin the U\nS\nSection 337’s reach is international and its adjudication is fast, making the ITC a compelling venue\nIn this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.\nConflict Minerals\nthe\nsmartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the\nS\nE\nC\nimplementing\nDodd-Frank\nrequiring all publicly traded companies to disclose\nuse of\nminerals associated with conflict and used in consumer goods\nFailure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors\nAnother “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law\nIn theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U\nS\nEnvironmental Degradation\nThe Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife\nany plant\nor\nprohibited wildlife species\ntaken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law\n“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”\nCollectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.\nA Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price\ncompanies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties\nor of importing lumber harvested in protected forests.\nIf a complainant can establish a\ncase\nunfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry\nthe ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place",
"role": "user"
},
{
"content": "I\nT\nC\neffective weapon\nunlawful trade practices\nrange of unfair acts\nprofoundly broad\nalready\ntrade secrets\ndesignation of origin\nlabeling\nad\ns\nCopyright\nantitrust\nexclusion order\nmarketing\ndistributing\nselling\nU\nS\nfast\ncompelling venue\nappeal of the U.S. marketplace\nfairer\nmarkets\nConflict Minerals\nmilitias\nS\nE\nC\nDodd-Frank\nrequir\ndisclose\nuse of\nminerals associated with conflict\nconsumer goods\ncomply\n337 action\nforeign sovereign’s law\nreach companies trading in certain minerals mined in Congo\nCongolese compliance\nU\nS\nEnvironmental Degradation\nLacey Act\ndemonstrating unfairness\nunlawful\nunfair\nfish\nwildlife\nplant\nprohibited wildlife species\nforeign law\nSection 337 investigation\nviolation of the Lacey Act\nflouted U.S. or foreign environmental law\nseafood\nlumber\nexcellent forum\nmore equitable place",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes.
Highlight Ratio: 0.2 | Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against
unlawful trade practices
Although Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language
is profoundly broad
already construed
to include misappropriation of trade secrets, false designation of origin, false labeling
false advertising, violation of the Digital Millennium Copyright Act and antitrust
Section 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling
in the U
S
Section 337’s reach is international and its adjudication is fast, making the ITC a compelling venue
In this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.
Conflict Minerals
the
smartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the
S
E
C
implementing
Dodd-Frank
requiring all publicly traded companies to disclose
use of
minerals associated with conflict and used in consumer goods
Failure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors
Another “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law
In theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U
S
Environmental Degradation
The Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife
any plant
or
prohibited wildlife species
taken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law
“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”
Collectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.
A Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price
companies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties
or of importing lumber harvested in protected forests.
If a complainant can establish a
case
unfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry
the ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place | the
I
T
C
could be
effective
against
unlawful
practices
a range
profoundly broad
already
include
trade secrets
designation of origin
labeling
ad
s
Copyright
and antitrust
rewards
complainants with
exclusion
can leverage the appeal of the U.S. marketplace to promote fairer
markets
The Lacey Act is
powerful
for demonstrating unfairness
makes it unlawful — and thus unfair
“to import
any fish
wildlife
plant
sold in violation
competitors
flouted U.S. or foreign environmental law for
bringing a commodity
at an artificially low price
the ITC is an excellent forum
to
make the world
more equitable | 256 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: That decimates the ITC, preventing its use for Lacey Act enforcement. That solves global environmental crimes.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "Section 337 of the Tariff Act of 1930 — a trade statute administered by the U.S. International Trade Commission — could be an effective weapon against\nunlawful trade practices\nAlthough Section 337 is best known as an intellectual property enforcement mechanism, the statute exists to protect U.S. industries and the American public from a range of unfair acts. The language\nis profoundly broad\nalready construed\nto include misappropriation of trade secrets, false designation of origin, false labeling\nfalse advertising, violation of the Digital Millennium Copyright Act and antitrust\nSection 337 rewards successful complainants with a powerful remedy: an exclusion order blocking the importation of goods associated with the unfairness. The ITC may also issue cease and desist orders that prohibit respondents from marketing, distributing or selling\nin the U\nS\nSection 337’s reach is international and its adjudication is fast, making the ITC a compelling venue\nIn this way, Section 337 can leverage the appeal of the U.S. marketplace to promote fairer and freer markets worldwide.\nConflict Minerals\nthe\nsmartphone or gaming system may contain raw materials mined by militias in war-torn countries. In August 2012, the\nS\nE\nC\nimplementing\nDodd-Frank\nrequiring all publicly traded companies to disclose\nuse of\nminerals associated with conflict and used in consumer goods\nFailure to comply could give rise to a Section 337 action, to the extent companies save money by not complying and, thus, are able to import and sell products to U.S. consumers at prices below those of their competitors\nAnother “unfair act” that could potentially trigger this type of case is the acquisition and use of conflict minerals in derogation of a foreign sovereign’s law\nIn theory, Section 337 could reach companies trading in certain minerals mined in Congo without performing the required Congolese compliance, where those activities result in an importation into the U\nS\nEnvironmental Degradation\nThe Lacey Act is a powerful weapon for demonstrating unfairness related to importation. Incredibly broad, the Lacey Act makes it unlawful — and thus unfair when undertaken for business advantage — “to import, export, transport, sell, receive, acquire, or purchase in interstate or foreign commerce ... any fish or wildlife\nany plant\nor\nprohibited wildlife species\ntaken, possessed, transported, or sold in violation of any law or regulation of any state or in violation of any foreign law\n“Taken” encompasses “captured, killed, or collected and, with respect to a plant, also means harvested, cut, logged, or removed.”\nCollectively, these terms encompass an enormous number of product lines of the Harmonized Tariff Schedule of the United States.\nA Section 337 investigation could be premised on the unfair act of importing an article in violation of the Lacey Act. Such a case would likely involve an allegation that the complainant’s competitors have flouted U.S. or foreign environmental law for the purpose of bringing a commodity to market at an artificially low price\ncompanies could be accused of importing seafood caught in prohibited waters or in violation of conservation treaties\nor of importing lumber harvested in protected forests.\nIf a complainant can establish a\ncase\nunfair acts or methods of competition, nexus between unfairness and importation and injury to a domestic industry\nthe ITC is an excellent forum in which to challenge harmful trade practices and, hopefully, make the world a more equitable place",
"role": "user"
},
{
"content": "the\nI\nT\nC\ncould be\neffective\nagainst\nunlawful\npractices\na range\nprofoundly broad\nalready\ninclude\ntrade secrets\ndesignation of origin\nlabeling\nad\ns\nCopyright\nand antitrust\nrewards\ncomplainants with\nexclusion\ncan leverage the appeal of the U.S. marketplace to promote fairer\nmarkets\nThe Lacey Act is\npowerful\nfor demonstrating unfairness\nmakes it unlawful — and thus unfair\n“to import\nany fish\nwildlife\nplant\nsold in violation\ncompetitors\nflouted U.S. or foreign environmental law for\nbringing a commodity\nat an artificially low price\nthe ITC is an excellent forum\nto\nmake the world\nmore equitable",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing. | The random manner in which civil-forfeiture taxes are levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in the commission of drug offenses. This group is politically unorganized and, hence, generally cannot adequately defend itself against more cohesively organized lobbies—such as law-enforcement agencies—who clamor for liberalized civil-forfeiture powers. Property owners at risk of forfeiture remain unorganized because they do not know who they are before having their properties confiscated under civil-forfeiture statutes. The potential loss of property due to forfeiture is typically a one-time, low-probability event for each property owner. Thus, they have little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes.[12]
The proceeds from civil-forfeiture actions thus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement.[13] Because the general population of taxpayers does not feel the full cost of government as readily as it would if government were financed entirely by direct tax payments, the result is excessively large government in general, and overly aggressive drug-law enforcement in particular.[14]
CIVIL FORFEITURE AND THE NONOPTIMALITY OF LAW ENFORCEMENT
In the previous section, we showed the biases civil forfeiture creates in the legislature; these biases result from exploitation of a revenue source that is at least partially hidden from general taxpayers. In this section, we explore the consequences of civil forfeiture for decisions made in the executive branch, where law-enforcement agencies get to keep all or part of the proceeds from such seizures.
As a preliminary step, however, we must first review some basic economics of law enforcement. Law enforcement, like other goods, is scarce. Resources used to enforce laws have alternative valuable uses, such as building more bridges and highways, providing more education, staffing larger social-service agencies, or keeping taxes lower so that private citizens can individually make their own resource allocation decisions. After some point, the gains from using additional resources to police against crime—even heinous crimes such as murder and rape—become smaller than the gains available from using these same resources elsewhere. Thus, completely wiping out all criminal activity would not be socially desirable.
As with all economic choices, the optimal amount of law enforcement occurs when the marginal benefit from an additional unit of enforcement just equals the marginal cost of the resources used to produce that additional enforcement. And what is true for law enforcement in general holds also for choices among alternative areas of law-enforcement activity. Each type of crime fighting entails costs. Devoting more crime-fighting resources to vice offenses means that fewer resources are available to police against murder and other illegal activities.[15]
Figure 1 depicts the costs and benefits of enforcing drug prohibition. In the top panel, the horizontal axis measures the quantity of drug-law enforcement, while the vertical axis shows the dollar value of the costs and benefits of such enforcement. The marginal-benefit curve (MB) shows the gains to society of each additional unit of drug-crime enforcement; the marginal-cost curve (MC) shows the costs of the additional units of enforcement. Enforcement beyond E* is excessive because the benefits from further reductions in drug crime are worth less to society than are all the goods and services sacrificed to generate these further reductions. Similarly, enforcement less than E* is suboptimal because the gains from additional enforcement then exceed the costs. Therefore, any institutional arrangement affecting law enforcement should be assessed by how likely it is to encourage the optimal amount of enforcement.
Civil forfeiture works against achieving the optimal amount of law enforcement. The reason is that civil forfeiture reduces the incentives law-enforcement officials have to attain E* by giving law-enforcement agents a disproportionate stake in the enforcement of drug prohibitions.
The lower part of Figure 1 shows the totals of civil-forfeiture revenues that law-enforcement agents retain as funds for their agencies.[16] These revenues can be depicted using the familiar “Laffer curve,” which shows the relationship between tax rates and tax receipts. If the agency engages in no drug-crime enforcement, it will gain no revenues from drug-related civil-forfeiture actions.
[[FIGURE OMITTED]]
This result is shown at point O. At the other extreme, if the agency enforces drug laws with such vigor as to eliminate illegal drug operations totally, the amount of civil-forfeiture revenues will also be zero. With no drug crime, there is no opportunity for drug-related civil forfeitures. This result is shown at point B. Between points 0 and B, however, civil-forfeiture revenues are positive. Civil-forfeiture revenues increase as drug-law enforcement is expanded from point 0 (no enforcement) toward greater enforcement. After some point, however, a greater intensity of enforcement reduces the number of drug offenses to such an extent that the dollar amounts available from seizures will also be reduced.
Revenues from seizures are maximized at EF, the “optimal” level of enforcement for law-enforcement agencies allowed to keep proceeds from civil-forfeiture seizures. Law-enforcement agencies will determine the intensity of their drug-enforcement efforts—and, by implication, the intensity of their efforts to enforce laws against nondrug crimes—by how such efforts affect their civil-forfeiture revenues. The revenue effects, rather than social-welfare consequences, of drug-crime enforcement determine the extent of police efforts in attacking various kinds of criminal behaviors.
No necessary correlation exists between EF (revenue-maximizing enforcement) and E* (socially optimal enforcement). EF could be to the right or to the left of E*; only by chance will EF be at the same level of enforcement as E*. Allowing law-enforcement agencies to retain proceeds from civil forfeitures affords agencies discretion over their budgets. As a consequence, agencies supply either sub or supraoptimal amounts of drug enforcement, thus leaving society worse off.
An alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement. With their budgets set by legislatures, law-enforcement agencies then allocate their monies among different types of activities, for example, drug-law enforcement versus policing against burglaries. In this scenario, no one type of activity has a “revenue advantage” over others in attracting law-enforcement resources.[17] The heads of law-enforcement agencies will thus make more rational enforcement decisions.
The legislature is the most appropriate branch of government for selecting the budget for law-enforcement agencies. The legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any of the specialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or for social-welfare programs or for any other of the multitude of items on which governments spend money. The economics of public choice make clear that legislatures are hardly likely to tax, spend, and regulate in ways that maximize society’s well-being.[18] Even so, legislatures are better placed to rationally allocate resources extracted from civil-forfeiture proceeds than are individual enforcement agencies.[19]
Can we be more precise and say whether the current civil-forfeiture regime causes over or underenforcement? We believe that overenforcement of drug laws is the more likely result. Overenforcement is more likely because of the victimless nature of drug use.[20] While all but a small handful of sociopaths prefer effective enforcement of laws curbing violent or fraudulent behaviors aimed at nonconsenting parties, it is much less obvious that a large number of Americans approve of the war on drugs. In 1990, in the midst of drug-war hysteria, the Drug Policy Foundation discovered that more than one in three Americans (36 percent) support legalization.[21] And more recently, in January 1994, NBC Nightly News conducted a telephone call-in poll asking callers if drugs should be legalized. A majority (52.2 percent) favored legalization.[22]
As popular support for a particular law decreases, it is plausible to assume that the social benefits of enforcing the law also decrease, especially in the case of victimless crimes. In Figure 1, this fall is shown by a marginal-benefit curve (MB’) closer to the origin than MB. The socially optimal level of enforcement falls as well (to E*’). Naturally, EF is more likely to be to the right of E* the further to the left is E*; that is, the weaker the consensus for outlawing a particular behavior, the more likely it is that legal prohibitions will be excessively enforced if enforcement agents derive personal benefits from their enforcement efforts. If no consensus exists for drug prohibition, then civil forfeiture probably leads to overenforcement of drug laws.
THE LEGAL PEDIGREE OF CIVIL FORFEITURE
The constitutionality of modern civil-forfeiture statutes does not depend on their efficiency. Many ill-thought statutes remain on the books with no one questioning their constitutionality. Nevertheless, there are solid arguments tending to show that most civil forfeitures under the Drug Act are unconstitutional. In this section, we briefly review the history of civil forfeiture, emphasizing germane Supreme Court decisions. After many years of deference to legislative choices in the realm of civil forfeiture, the Court now seems to be coming around, albeit slowly, to the view that civil forfeiture poses a grave and unconstitutional threat to American liberties.
The roots of American forfeiture law are found in England. The predominant kind of forfeiture in England was not civil forfeiture at all; it was criminal forfeiture (which remains on the statute books). Criminal forfeiture occurs only after a person has been convicted of a criminal offense. The convict’s property is seized and liquidated by government, following a criminal conviction, as part of the punishment visited upon those proven guilty.
The strict procedural requirements of the criminal law (e.g., trial by jury, a presumption of innocence, the privilege against self-incrimination, and protection against double jeopardy) help thwart unjustified seizures by government. With a criminal conviction as a necessary prerequisite to forfeiture, criminal forfeiture affords greater assurance that people paying fines or restitution to government are actually guilty of antisocial behaviors. Moreover, the Eighth Amendment’s protection against excessive fines helps keep property loss under criminal forfeiture from growing out of proportion to the seriousness of the crime.
When government prosecutes someone for a crime, the suit is said to be in personam, which is legal jargon for a suit against an identifiable person. The individual accused of criminal wrongdoing can be punished only after government wins a conviction in court. Criminal forfeitures are always in personam prosecutions.
Contrast criminal forfeiture with civil forfeiture. In civil forfeiture, the government sues the offending items rather than any human being. In legal jargon, the government proceeds in rem; that is, the government sues the thing (rather than the thing’s owner).[23] All civil-forfeiture suits are in rem suits. For example, if Mr. Smith’s house at 159 Elm Street is thought by the police to have been the site of a cocaine sale, then the government will sue the house at 159 Elm Street, charging it with wrongdoing. Because criminal-law protections are available only to people, the defendant in this hypothetical civil case—formally, the house at 159 Elm Street—can be found liable for wrongdoing and confiscated and liquidated by government without the usual constitutional formalities. If Mr. Smith complains that he has been deprived of his property without ever having been charged with a crime, the government resorts to the legal fiction that the house was punished, not Mr. Smith, so that Mr. Smith has no cause to complain.
Understanding the origin of civil forfeiture (i.e., of in rem legal proceedings) will shed light on contemporary practice. In pre-Civil War America, as under English common law, in rem forfeiture proceedings were used in only two related contexts: (1) to enforce customs law (customs duties being the main source of federal revenue), and (2) to enforce admiralty laws (such as laws against piracy). If the government found smuggled goods, it could seize and liquidate them for revenue. The government was not obliged to criminally convict the owner of the goods prior to forfeiture. Likewise, a seized pirate ship was forfeitable without securing a criminal conviction of the vessel’s owner. Practical concerns justified dropping the requirement of criminal conviction of property owners in these cases. Owners were likely to be in foreign lands, beyond the jurisdiction of domestic courts. Rather than let these defendants break the law without penalty, the government cleverly devised the legal fiction of in rem civil suits to govern those classes of cases in which defendants were not likely to be within the jurisdiction of domestic courts.
Historically, in England and in the United States until the second half of the 19th century, civil-forfeiture suits were limited to the narrow class of cases (having to do with customs or admiralty law) in which the overwhelming impracticality of gaining in personam jurisdiction over property owners dictated an alternative means of penalizing wrongdoers. Augmenting government revenues with confiscated valuable properties was not the rationale for the civil (in rem) proceedings in these special cases; rather, even-handed enforcement of the customs and admiralty laws required proceeding without the owners’ consent or participation. Under common law, real property was subject to forfeiture only after a criminal conviction of the owner—and the first U.S. Congress, by statute, abolished criminal forfeiture altogether. With the Civil War, however, came the beginning of government efforts to use civil-forfeiture proceedings as a general method of avoiding the procedural requirements of the Constitution. Rather than limit civil forfeiture to those instances when it was practically impossible to gain jurisdiction over an owner, civil forfeiture came to be used as a device by which the government sought to evade constitutional restrictions.
Beginning in August 1861, Congress enacted the Confiscation Acts authorizing the seizure of property owned by Confederates and those aiding the rebellion.[24] These Acts permitted in rem proceedings for the confiscation of both personal and real property. The Acts’ coverage of real property is particularly noteworthy. Until the Civil War, civil forfeiture was limited to personal property (mainly ships or contraband goods); it did not extend to real property. The reason it did not is a straightforward one: real property (unlike personal property) cannot be contraband. Moreover, the owners of real properties used in unlawful ways are more likely to be within the jurisdiction of domestic courts than are the owners of the personal properties traditionally subject to civil forfeiture. Unlike personal property, real property cannot flee during the time it takes to find and establish jurisdiction over the owner.
But wartime exigencies notoriously serve as pretexts for expansion of government power.[24] The Supreme Court upheld the Confiscation Acts as a legitimate exercise of the war power, rejecting the argument that the Acts violated the Fifth and Sixth Amendments. The Court also dismissed the contention that due process requires a criminal conviction before forfeiture.[26] By labeling such punishments “civil,” despite their obviously punitive nature, the government succeeded in nullifying constitutional protections against criminal punishment. The war power served as the excuse for doing so.
The upholding of the Confiscation Acts encouraged Congress to expand in rem forfeiture further beyond its traditional domain—customs and the admiralty jurisdiction—to enforce revenue provisions unrelated to the maritime trade or to the war power.[27] In Dobbins’s Distillery v. United States, the Court upheld the civil forfeiture of a distillery, and the real property upon which it stood, for liquortax violations.[28] The tenant on the property was accused of defrauding the government of the excise taxes due on the liquor distilled there.[29] The Court held that it was not “necessary that the owner of the property should have knowledge that the tenant was committing fraud on the public revenue, in order that the information of forfeiture should be maintained.”[30]
The Court reasoned that the proceeding was civil in nature, and that any “conviction of the wrong-doer must be obtained, if at all, in another and wholly independent proceeding.”[31] But no precedent is cited for the proposition that a forfeiture of real property, as opposed to the personal property traditionally subject to forfeiture in in rem proceedings, can be accomplished without a prior criminal conviction;[32] indeed, the opinion nowhere indicates that the Court even considered that real property might be distinguished from personal property. The Court affirmed the judgment of forfeiture, despite the absence of wrongdoing by the owner, because “the offence...is attached primarily to the distillery, and the real and personal property used in connection with the same, without any regard whatsoever to the personal misconduct or responsibility of the owner.”[33]
The Confiscation Acts cases, along with Dobbins’s Distillery, thus mark a watershed in the Court’s treatment of civil-forfeiture statutes. The Court has opened the door so that any kind of property is subject to civil forfeiture as long as the government labels its enforcement actions “civil.” The initial limited rationale for allowing forfeitures in civil rather than criminal proceedings was ignored or forgotten. Civil forfeiture became a generally available weapon for use by government against wrongdoers, and the Constitution’s guarantee of certain rights and privileges to the criminally accused would not block government’s way if the legislature took care to label a particular forfeiture action “civil” rather than “criminal.”
During the next century, the Court never seemed quite sure what to do with civil-forfeiture statutes. Although the genie of civil forfeiture was out of the bottle, every now and then the Court would declare a specific application of this or that civil-forfeiture statute unconstitutional. But such declarations never amounted to any wholesale scaling back of the use of civil forfeiture. For example, in Coffey v. United States, the Court held that a prior acquittal of the crime of violating internal-revenue laws barred a subsequent civil-forfeiture suit based on actions for which the defendant was earlier found innocent;[34] that is, civil-forfeiture defendants were thus extended the constitutional protection against double jeopardy. And later in the same term, the Court held in Boyd v. United States that both the Fourth and Fifth Amendments protected the defendant against compulsory production of his private books and papers in a civil-forfeiture proceeding. The Court declared the proceedings to be “quasi-criminal,” reasoning “that proceedings instituted for the purpose of declaring the forfeiture of a man’s property by reasons of offences committed by him, though they may be civil in form, are in their nature criminal.”[35] Boyd thus promised that the Court would no longer be duped into treating the forfeiture of property as a civil matter simply because the legislature labeled it “civil.”
Boyd’s promise was quickly abandoned. Subsequent decisions served to limit Boyd’s reach to the narrow Fourth and Fifth Amendment grounds on which it was decided. The Constitution’s other protections of the criminally accused were held not to apply in civil-forfeiture actions because these actions are “civil” and not “criminal.” For example, in United States v. Zucker, the Court held that the right of a criminal defendant to confront his accusers does not apply to civil forfeiture.[36] Also, proof beyond a reasonable doubt was found by the Court not to be required in civil-forfeiture proceedings.[37] Finally, the Court declared that there is no constitutional requirement that the jury find the defendant guilty in civil-forfeiture proceedings; judges are free to direct verdicts in favor of government.[38]
Ominously, Coffey’s protection against double jeopardy in the civil-forfeiture context is no longer good law. In 1984, the Court declared that the double-jeopardy clause does not bar “a civil, remedial forfeiture proceeding initiated following an acquittal on related criminal charges. To the extent that Coffey v. United States suggests otherwise, it is hereby overruled.”[39] And Boyd has been all but overruled. In United States v. Ward, the Court accurately noted that it “has declined...to give full scope to the reasoning and dicta in Boyd.”[40]
The most notable recent refusal of the Court to restrain civil forfeiture is the 1974 case of Calero-Toledo v. Pearson Yacht Leasing Co.[41] In Calero-Toledo, a yacht worth $20,000 was leased by Pearson Yacht Leasing to a party who subsequently smoked or transported marijuana on the yacht. Upon the discovery of two marijuana cigarettes on the vessel, local authorities—under a Puerto Rican civil-forfeiture statute—seized the yacht without a prior judicial hearing or without notice being given to Pearson. All parties admitted that Pearson had no knowledge that marijuana was used or carried on the vessel. However, the Supreme Court upheld the forfeiture on the grounds that Pearson had failed to prove that it had done “all that reasonably could be expected to prevent the proscribed use of [its] property.”[42] The Court ruled that Pearson’s right to due process had not been violated.
In effect, Calero-Toledo reversed the presumption of innocence that is the Constitutional right of persons subject to criminal prosecution. The government was presumed to have behaved properly, and the burden of proving that the government overstepped its authority by seizing a vessel whose owner had done no wrong was foisted upon the owner. If the Puerto Rican statute had been labeled “criminal” rather than “civil,” of course, this reversal of the presumption of innocence would have been held unconstitutional. But, as it had so often in the past, the Court simply trusted the legislature to choose the correct labels for its actions: the Puerto Rican legislature said that this action was civil, so it must be civil, and thus, constitutional protections available to the criminally accused are inapplicable.
Fortunately, a trend against unchecked use of civil forfeiture is just now emerging in the Court. During two recent terms (1992–93 and 1993–94), the Court heard four civil-forfeiture cases to which the United States government was a party. The government lost all four.[43] Each of these cases exhibits intensified judicial skepticism of the use of civil forfeiture in fighting the war on drugs. We focus here on the last two of these cases, Austin and James Daniel Good.
In Austin, the government sought the civil forfeiture of Richard Austin’s mobile home and auto-body shop after he was convicted of distributing two ounces of cocaine from the shop, having brought the cocaine from his mobile home to consummate a prearranged sale. Both the trial court and the appellate court rejected Austin’s argument that the forfeiture would violate his Eighth Amendment right to be protected from excessive fines. Austin appealed to the Supreme Court.
The government urged that civil forfeiture of real property is not a criminal punishment and, therefore, the government can-not be held to the strict requirements of criminal procedure when it confiscates real property. The Court rejected this argument, unanimously reversing the lower court. The Court found that “forfeiture generally and statutory in rem forfeiture in particular historically have been understood, at least in part, as punishment,”[44] thus repudiating the government’s assertion that the Drug Act’s forfeiture provisions are not punitive. The Court concluded that the forfeiture was a “payment to a sovereign as punishment for some offense...subject to the limitations of the Eighth Amendment’s Excessive Fines Clause.”[45]
The case of James Daniel Good also represents a step forward in restoring liberties trampled underfoot by a government motivated by civil-forfeiture revenues. Four years after James Daniel Good was convicted, sentenced, and fined in a Hawaii state court for growing marijuana, the United States government seized his house and the four acres of land upon which it sits. Good received no prior notice of the seizure. Although the trial court upheld the seizure, the court of appeals reversed it, finding that the government violated Mr. Good’s right to due process of law by not affording him notice and a hearing prior to the seizure.
Arguing before the Supreme Court, the government insisted that the due-process clause does not apply when the government seizes property for law-enforcement purposes.[46] The Court did not buy this claim. In James Daniel Good, the Court reasoned that the government’s asserted interests in seizing property—obtaining jurisdiction and preventing the loss or destruction of the property—were slight in the case of real property. The Court ruled that due process requires the government to give notice and a hearing to owners of real property before seizures because of (1) the high risk of error created by allowing government to seize property without adhering to the requirements of due process of law, and (2) the substantial interests citizens have in being secure in their homes.
Thus, in just two terms of the Court, substantial constitutional constraints have been placed on government’s exercise of civil-forfeiture authority. Civil forfeiture remains, however, a source of revenue for government as it wages its war on drugs. For example, although James Daniel Good requires notice and a hearing prior to seizure of real property, the burden remains upon property owners to prove their innocence. Failure to prove innocence at a preseizure hearing means that government may seize the property and liquidate it for revenue purposes. We believe that more can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture. Indeed, as we read the Constitution, more must be done.
At a minimum, civil forfeitures of real property should all be declared unconstitutional. The common law of England and the United States at the time of the Constitution’s ratification did not provide for civil forfeiture of real properties; such forfeitures could be accomplished only following criminal convictions of the owners of real property. Unlike captured pirate ships and contraband goods, there is no chance that real property can stealthily escape jurisdiction before its owner is convicted. Consequently, the pragmatic rationale for allowing in rem seizures of personal property does not apply to real property. Because in rem forfeiture of real property was unheard of at the time the Constitution was ratified, the framers and ratifiers of the Constitution cannot plausibly be supposed to have intended to permit such forfeitures. Thus, any forfeitures of real property sought by government should be part of the explicit punishment sought in criminal proceedings.
Given the abuses that have marked civil forfeiture under the Drug Act, the Court might consider it wise to adopt an even more stringent approach to civil forfeiture. Under such an approach, the government would be barred from using civil procedures for the forfeiture of any property—real or personal—whenever (1) the property is forfeitable to the government because of its use in criminal wrongdoing, and (2) the government cannot prove that the property owner is beyond the personal jurisdiction of domestic courts. If the property owner is within the jurisdiction of domestic courts, then he or she should be duly accused, indicted, tried, and convicted before his or her property is forfeited to government. The protections of the criminal process are essential to safeguard innocent parties from government overreaching. If the owner is outside of domestic jurisdiction, the government must then, in good faith, attempt to notify the owner that his or her property temporarily (pending the outcome of a criminal trial) has been seized. If the owner then appears to wage a defense, a criminal conviction should be necessary before final forfeiture. In contrast, if the owner refuses to attend the trial, the property can then be forfeited in a civil action. In either case, however, the burden should be upon the government to prove wrongdoing by the owner, and not (as it is now) upon the owner to prove his or her innocence of any wrongdoing.
Given that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive in purpose?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture), the court should then ask, “Is the property contraband, or is it a legal good?” If it is contra-band, then the government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner.[47]
When a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final. If the owner of the property is outside the court’s jurisdiction, then fair notice to the defendant of the charges is necessary, as is a fair opportunity for defense. If the defendant fails to show for his or her trial after being notified (or after a reasonable attempt to notify has been made), then a conviction obtained in absentia permits forfeiture to the government of the defendant’s property. The value of such forfeitures would still be limited, of course, by the excessive-fines clause of the Eighth Amendment. | The random manner in which civil-forfeiture
are levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in
offenses. This group is politically unorganized and
generally cannot adequately defend itself against
lobbies
such as law-enforcement agencies
who clamor for
civil-forfeiture powers. Property owners
have little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes
The proceeds
thus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement
Because the general population of taxpayers does not feel the full cost of government
the result is excessively
aggressive drug-law enforcement
biases result from exploitation of a revenue source that is
hidden from general taxpayers
An alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement
In this scenario
no one type of activity has a “revenue advantage” over others in attracting law-enforcement resources
heads of law-enforcement agencies will thus make more rational enforcement decisions
The legislature is
most appropriate
for selecting the budget for law-enforcement agencies
The legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any
specialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or
welfare
legislatures are better placed to rationally allocate resources
than are individual enforcement agencies
Civil forfeiture remains
however, a source of revenue for government as it wages its war on drugs
more can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture
more must be done
The protections of the criminal process are essential to safeguard innocent parties from
overreach
Given that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive
?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture
the court should then ask, “Is the property contraband
?” If it is
the government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner
When a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final | 257 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing.",
"role": "system"
},
{
"content": "The random manner in which civil-forfeiture taxes are levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in the commission of drug offenses. This group is politically unorganized and, hence, generally cannot adequately defend itself against more cohesively organized lobbies—such as law-enforcement agencies—who clamor for liberalized civil-forfeiture powers. Property owners at risk of forfeiture remain unorganized because they do not know who they are before having their properties confiscated under civil-forfeiture statutes. The potential loss of property due to forfeiture is typically a one-time, low-probability event for each property owner. Thus, they have little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes.[12]\nThe proceeds from civil-forfeiture actions thus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement.[13] Because the general population of taxpayers does not feel the full cost of government as readily as it would if government were financed entirely by direct tax payments, the result is excessively large government in general, and overly aggressive drug-law enforcement in particular.[14]\nCIVIL FORFEITURE AND THE NONOPTIMALITY OF LAW ENFORCEMENT\nIn the previous section, we showed the biases civil forfeiture creates in the legislature; these biases result from exploitation of a revenue source that is at least partially hidden from general taxpayers. In this section, we explore the consequences of civil forfeiture for decisions made in the executive branch, where law-enforcement agencies get to keep all or part of the proceeds from such seizures.\nAs a preliminary step, however, we must first review some basic economics of law enforcement. Law enforcement, like other goods, is scarce. Resources used to enforce laws have alternative valuable uses, such as building more bridges and highways, providing more education, staffing larger social-service agencies, or keeping taxes lower so that private citizens can individually make their own resource allocation decisions. After some point, the gains from using additional resources to police against crime—even heinous crimes such as murder and rape—become smaller than the gains available from using these same resources elsewhere. Thus, completely wiping out all criminal activity would not be socially desirable.\nAs with all economic choices, the optimal amount of law enforcement occurs when the marginal benefit from an additional unit of enforcement just equals the marginal cost of the resources used to produce that additional enforcement. And what is true for law enforcement in general holds also for choices among alternative areas of law-enforcement activity. Each type of crime fighting entails costs. Devoting more crime-fighting resources to vice offenses means that fewer resources are available to police against murder and other illegal activities.[15]\nFigure 1 depicts the costs and benefits of enforcing drug prohibition. In the top panel, the horizontal axis measures the quantity of drug-law enforcement, while the vertical axis shows the dollar value of the costs and benefits of such enforcement. The marginal-benefit curve (MB) shows the gains to society of each additional unit of drug-crime enforcement; the marginal-cost curve (MC) shows the costs of the additional units of enforcement. Enforcement beyond E* is excessive because the benefits from further reductions in drug crime are worth less to society than are all the goods and services sacrificed to generate these further reductions. Similarly, enforcement less than E* is suboptimal because the gains from additional enforcement then exceed the costs. Therefore, any institutional arrangement affecting law enforcement should be assessed by how likely it is to encourage the optimal amount of enforcement.\nCivil forfeiture works against achieving the optimal amount of law enforcement. The reason is that civil forfeiture reduces the incentives law-enforcement officials have to attain E* by giving law-enforcement agents a disproportionate stake in the enforcement of drug prohibitions.\nThe lower part of Figure 1 shows the totals of civil-forfeiture revenues that law-enforcement agents retain as funds for their agencies.[16] These revenues can be depicted using the familiar “Laffer curve,” which shows the relationship between tax rates and tax receipts. If the agency engages in no drug-crime enforcement, it will gain no revenues from drug-related civil-forfeiture actions.\n[[FIGURE OMITTED]]\nThis result is shown at point O. At the other extreme, if the agency enforces drug laws with such vigor as to eliminate illegal drug operations totally, the amount of civil-forfeiture revenues will also be zero. With no drug crime, there is no opportunity for drug-related civil forfeitures. This result is shown at point B. Between points 0 and B, however, civil-forfeiture revenues are positive. Civil-forfeiture revenues increase as drug-law enforcement is expanded from point 0 (no enforcement) toward greater enforcement. After some point, however, a greater intensity of enforcement reduces the number of drug offenses to such an extent that the dollar amounts available from seizures will also be reduced.\nRevenues from seizures are maximized at EF, the “optimal” level of enforcement for law-enforcement agencies allowed to keep proceeds from civil-forfeiture seizures. Law-enforcement agencies will determine the intensity of their drug-enforcement efforts—and, by implication, the intensity of their efforts to enforce laws against nondrug crimes—by how such efforts affect their civil-forfeiture revenues. The revenue effects, rather than social-welfare consequences, of drug-crime enforcement determine the extent of police efforts in attacking various kinds of criminal behaviors.\nNo necessary correlation exists between EF (revenue-maximizing enforcement) and E* (socially optimal enforcement). EF could be to the right or to the left of E*; only by chance will EF be at the same level of enforcement as E*. Allowing law-enforcement agencies to retain proceeds from civil forfeitures affords agencies discretion over their budgets. As a consequence, agencies supply either sub or supraoptimal amounts of drug enforcement, thus leaving society worse off.\nAn alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement. With their budgets set by legislatures, law-enforcement agencies then allocate their monies among different types of activities, for example, drug-law enforcement versus policing against burglaries. In this scenario, no one type of activity has a “revenue advantage” over others in attracting law-enforcement resources.[17] The heads of law-enforcement agencies will thus make more rational enforcement decisions.\nThe legislature is the most appropriate branch of government for selecting the budget for law-enforcement agencies. The legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any of the specialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or for social-welfare programs or for any other of the multitude of items on which governments spend money. The economics of public choice make clear that legislatures are hardly likely to tax, spend, and regulate in ways that maximize society’s well-being.[18] Even so, legislatures are better placed to rationally allocate resources extracted from civil-forfeiture proceeds than are individual enforcement agencies.[19]\nCan we be more precise and say whether the current civil-forfeiture regime causes over or underenforcement? We believe that overenforcement of drug laws is the more likely result. Overenforcement is more likely because of the victimless nature of drug use.[20] While all but a small handful of sociopaths prefer effective enforcement of laws curbing violent or fraudulent behaviors aimed at nonconsenting parties, it is much less obvious that a large number of Americans approve of the war on drugs. In 1990, in the midst of drug-war hysteria, the Drug Policy Foundation discovered that more than one in three Americans (36 percent) support legalization.[21] And more recently, in January 1994, NBC Nightly News conducted a telephone call-in poll asking callers if drugs should be legalized. A majority (52.2 percent) favored legalization.[22]\nAs popular support for a particular law decreases, it is plausible to assume that the social benefits of enforcing the law also decrease, especially in the case of victimless crimes. In Figure 1, this fall is shown by a marginal-benefit curve (MB’) closer to the origin than MB. The socially optimal level of enforcement falls as well (to E*’). Naturally, EF is more likely to be to the right of E* the further to the left is E*; that is, the weaker the consensus for outlawing a particular behavior, the more likely it is that legal prohibitions will be excessively enforced if enforcement agents derive personal benefits from their enforcement efforts. If no consensus exists for drug prohibition, then civil forfeiture probably leads to overenforcement of drug laws.\nTHE LEGAL PEDIGREE OF CIVIL FORFEITURE\nThe constitutionality of modern civil-forfeiture statutes does not depend on their efficiency. Many ill-thought statutes remain on the books with no one questioning their constitutionality. Nevertheless, there are solid arguments tending to show that most civil forfeitures under the Drug Act are unconstitutional. In this section, we briefly review the history of civil forfeiture, emphasizing germane Supreme Court decisions. After many years of deference to legislative choices in the realm of civil forfeiture, the Court now seems to be coming around, albeit slowly, to the view that civil forfeiture poses a grave and unconstitutional threat to American liberties.\nThe roots of American forfeiture law are found in England. The predominant kind of forfeiture in England was not civil forfeiture at all; it was criminal forfeiture (which remains on the statute books). Criminal forfeiture occurs only after a person has been convicted of a criminal offense. The convict’s property is seized and liquidated by government, following a criminal conviction, as part of the punishment visited upon those proven guilty.\nThe strict procedural requirements of the criminal law (e.g., trial by jury, a presumption of innocence, the privilege against self-incrimination, and protection against double jeopardy) help thwart unjustified seizures by government. With a criminal conviction as a necessary prerequisite to forfeiture, criminal forfeiture affords greater assurance that people paying fines or restitution to government are actually guilty of antisocial behaviors. Moreover, the Eighth Amendment’s protection against excessive fines helps keep property loss under criminal forfeiture from growing out of proportion to the seriousness of the crime.\nWhen government prosecutes someone for a crime, the suit is said to be in personam, which is legal jargon for a suit against an identifiable person. The individual accused of criminal wrongdoing can be punished only after government wins a conviction in court. Criminal forfeitures are always in personam prosecutions.\nContrast criminal forfeiture with civil forfeiture. In civil forfeiture, the government sues the offending items rather than any human being. In legal jargon, the government proceeds in rem; that is, the government sues the thing (rather than the thing’s owner).[23] All civil-forfeiture suits are in rem suits. For example, if Mr. Smith’s house at 159 Elm Street is thought by the police to have been the site of a cocaine sale, then the government will sue the house at 159 Elm Street, charging it with wrongdoing. Because criminal-law protections are available only to people, the defendant in this hypothetical civil case—formally, the house at 159 Elm Street—can be found liable for wrongdoing and confiscated and liquidated by government without the usual constitutional formalities. If Mr. Smith complains that he has been deprived of his property without ever having been charged with a crime, the government resorts to the legal fiction that the house was punished, not Mr. Smith, so that Mr. Smith has no cause to complain.\nUnderstanding the origin of civil forfeiture (i.e., of in rem legal proceedings) will shed light on contemporary practice. In pre-Civil War America, as under English common law, in rem forfeiture proceedings were used in only two related contexts: (1) to enforce customs law (customs duties being the main source of federal revenue), and (2) to enforce admiralty laws (such as laws against piracy). If the government found smuggled goods, it could seize and liquidate them for revenue. The government was not obliged to criminally convict the owner of the goods prior to forfeiture. Likewise, a seized pirate ship was forfeitable without securing a criminal conviction of the vessel’s owner. Practical concerns justified dropping the requirement of criminal conviction of property owners in these cases. Owners were likely to be in foreign lands, beyond the jurisdiction of domestic courts. Rather than let these defendants break the law without penalty, the government cleverly devised the legal fiction of in rem civil suits to govern those classes of cases in which defendants were not likely to be within the jurisdiction of domestic courts.\nHistorically, in England and in the United States until the second half of the 19th century, civil-forfeiture suits were limited to the narrow class of cases (having to do with customs or admiralty law) in which the overwhelming impracticality of gaining in personam jurisdiction over property owners dictated an alternative means of penalizing wrongdoers. Augmenting government revenues with confiscated valuable properties was not the rationale for the civil (in rem) proceedings in these special cases; rather, even-handed enforcement of the customs and admiralty laws required proceeding without the owners’ consent or participation. Under common law, real property was subject to forfeiture only after a criminal conviction of the owner—and the first U.S. Congress, by statute, abolished criminal forfeiture altogether. With the Civil War, however, came the beginning of government efforts to use civil-forfeiture proceedings as a general method of avoiding the procedural requirements of the Constitution. Rather than limit civil forfeiture to those instances when it was practically impossible to gain jurisdiction over an owner, civil forfeiture came to be used as a device by which the government sought to evade constitutional restrictions.\nBeginning in August 1861, Congress enacted the Confiscation Acts authorizing the seizure of property owned by Confederates and those aiding the rebellion.[24] These Acts permitted in rem proceedings for the confiscation of both personal and real property. The Acts’ coverage of real property is particularly noteworthy. Until the Civil War, civil forfeiture was limited to personal property (mainly ships or contraband goods); it did not extend to real property. The reason it did not is a straightforward one: real property (unlike personal property) cannot be contraband. Moreover, the owners of real properties used in unlawful ways are more likely to be within the jurisdiction of domestic courts than are the owners of the personal properties traditionally subject to civil forfeiture. Unlike personal property, real property cannot flee during the time it takes to find and establish jurisdiction over the owner.\nBut wartime exigencies notoriously serve as pretexts for expansion of government power.[24] The Supreme Court upheld the Confiscation Acts as a legitimate exercise of the war power, rejecting the argument that the Acts violated the Fifth and Sixth Amendments. The Court also dismissed the contention that due process requires a criminal conviction before forfeiture.[26] By labeling such punishments “civil,” despite their obviously punitive nature, the government succeeded in nullifying constitutional protections against criminal punishment. The war power served as the excuse for doing so.\nThe upholding of the Confiscation Acts encouraged Congress to expand in rem forfeiture further beyond its traditional domain—customs and the admiralty jurisdiction—to enforce revenue provisions unrelated to the maritime trade or to the war power.[27] In Dobbins’s Distillery v. United States, the Court upheld the civil forfeiture of a distillery, and the real property upon which it stood, for liquortax violations.[28] The tenant on the property was accused of defrauding the government of the excise taxes due on the liquor distilled there.[29] The Court held that it was not “necessary that the owner of the property should have knowledge that the tenant was committing fraud on the public revenue, in order that the information of forfeiture should be maintained.”[30]\nThe Court reasoned that the proceeding was civil in nature, and that any “conviction of the wrong-doer must be obtained, if at all, in another and wholly independent proceeding.”[31] But no precedent is cited for the proposition that a forfeiture of real property, as opposed to the personal property traditionally subject to forfeiture in in rem proceedings, can be accomplished without a prior criminal conviction;[32] indeed, the opinion nowhere indicates that the Court even considered that real property might be distinguished from personal property. The Court affirmed the judgment of forfeiture, despite the absence of wrongdoing by the owner, because “the offence...is attached primarily to the distillery, and the real and personal property used in connection with the same, without any regard whatsoever to the personal misconduct or responsibility of the owner.”[33]\nThe Confiscation Acts cases, along with Dobbins’s Distillery, thus mark a watershed in the Court’s treatment of civil-forfeiture statutes. The Court has opened the door so that any kind of property is subject to civil forfeiture as long as the government labels its enforcement actions “civil.” The initial limited rationale for allowing forfeitures in civil rather than criminal proceedings was ignored or forgotten. Civil forfeiture became a generally available weapon for use by government against wrongdoers, and the Constitution’s guarantee of certain rights and privileges to the criminally accused would not block government’s way if the legislature took care to label a particular forfeiture action “civil” rather than “criminal.”\nDuring the next century, the Court never seemed quite sure what to do with civil-forfeiture statutes. Although the genie of civil forfeiture was out of the bottle, every now and then the Court would declare a specific application of this or that civil-forfeiture statute unconstitutional. But such declarations never amounted to any wholesale scaling back of the use of civil forfeiture. For example, in Coffey v. United States, the Court held that a prior acquittal of the crime of violating internal-revenue laws barred a subsequent civil-forfeiture suit based on actions for which the defendant was earlier found innocent;[34] that is, civil-forfeiture defendants were thus extended the constitutional protection against double jeopardy. And later in the same term, the Court held in Boyd v. United States that both the Fourth and Fifth Amendments protected the defendant against compulsory production of his private books and papers in a civil-forfeiture proceeding. The Court declared the proceedings to be “quasi-criminal,” reasoning “that proceedings instituted for the purpose of declaring the forfeiture of a man’s property by reasons of offences committed by him, though they may be civil in form, are in their nature criminal.”[35] Boyd thus promised that the Court would no longer be duped into treating the forfeiture of property as a civil matter simply because the legislature labeled it “civil.”\nBoyd’s promise was quickly abandoned. Subsequent decisions served to limit Boyd’s reach to the narrow Fourth and Fifth Amendment grounds on which it was decided. The Constitution’s other protections of the criminally accused were held not to apply in civil-forfeiture actions because these actions are “civil” and not “criminal.” For example, in United States v. Zucker, the Court held that the right of a criminal defendant to confront his accusers does not apply to civil forfeiture.[36] Also, proof beyond a reasonable doubt was found by the Court not to be required in civil-forfeiture proceedings.[37] Finally, the Court declared that there is no constitutional requirement that the jury find the defendant guilty in civil-forfeiture proceedings; judges are free to direct verdicts in favor of government.[38]\nOminously, Coffey’s protection against double jeopardy in the civil-forfeiture context is no longer good law. In 1984, the Court declared that the double-jeopardy clause does not bar “a civil, remedial forfeiture proceeding initiated following an acquittal on related criminal charges. To the extent that Coffey v. United States suggests otherwise, it is hereby overruled.”[39] And Boyd has been all but overruled. In United States v. Ward, the Court accurately noted that it “has declined...to give full scope to the reasoning and dicta in Boyd.”[40]\nThe most notable recent refusal of the Court to restrain civil forfeiture is the 1974 case of Calero-Toledo v. Pearson Yacht Leasing Co.[41] In Calero-Toledo, a yacht worth $20,000 was leased by Pearson Yacht Leasing to a party who subsequently smoked or transported marijuana on the yacht. Upon the discovery of two marijuana cigarettes on the vessel, local authorities—under a Puerto Rican civil-forfeiture statute—seized the yacht without a prior judicial hearing or without notice being given to Pearson. All parties admitted that Pearson had no knowledge that marijuana was used or carried on the vessel. However, the Supreme Court upheld the forfeiture on the grounds that Pearson had failed to prove that it had done “all that reasonably could be expected to prevent the proscribed use of [its] property.”[42] The Court ruled that Pearson’s right to due process had not been violated.\nIn effect, Calero-Toledo reversed the presumption of innocence that is the Constitutional right of persons subject to criminal prosecution. The government was presumed to have behaved properly, and the burden of proving that the government overstepped its authority by seizing a vessel whose owner had done no wrong was foisted upon the owner. If the Puerto Rican statute had been labeled “criminal” rather than “civil,” of course, this reversal of the presumption of innocence would have been held unconstitutional. But, as it had so often in the past, the Court simply trusted the legislature to choose the correct labels for its actions: the Puerto Rican legislature said that this action was civil, so it must be civil, and thus, constitutional protections available to the criminally accused are inapplicable.\nFortunately, a trend against unchecked use of civil forfeiture is just now emerging in the Court. During two recent terms (1992–93 and 1993–94), the Court heard four civil-forfeiture cases to which the United States government was a party. The government lost all four.[43] Each of these cases exhibits intensified judicial skepticism of the use of civil forfeiture in fighting the war on drugs. We focus here on the last two of these cases, Austin and James Daniel Good.\nIn Austin, the government sought the civil forfeiture of Richard Austin’s mobile home and auto-body shop after he was convicted of distributing two ounces of cocaine from the shop, having brought the cocaine from his mobile home to consummate a prearranged sale. Both the trial court and the appellate court rejected Austin’s argument that the forfeiture would violate his Eighth Amendment right to be protected from excessive fines. Austin appealed to the Supreme Court.\nThe government urged that civil forfeiture of real property is not a criminal punishment and, therefore, the government can-not be held to the strict requirements of criminal procedure when it confiscates real property. The Court rejected this argument, unanimously reversing the lower court. The Court found that “forfeiture generally and statutory in rem forfeiture in particular historically have been understood, at least in part, as punishment,”[44] thus repudiating the government’s assertion that the Drug Act’s forfeiture provisions are not punitive. The Court concluded that the forfeiture was a “payment to a sovereign as punishment for some offense...subject to the limitations of the Eighth Amendment’s Excessive Fines Clause.”[45]\nThe case of James Daniel Good also represents a step forward in restoring liberties trampled underfoot by a government motivated by civil-forfeiture revenues. Four years after James Daniel Good was convicted, sentenced, and fined in a Hawaii state court for growing marijuana, the United States government seized his house and the four acres of land upon which it sits. Good received no prior notice of the seizure. Although the trial court upheld the seizure, the court of appeals reversed it, finding that the government violated Mr. Good’s right to due process of law by not affording him notice and a hearing prior to the seizure.\nArguing before the Supreme Court, the government insisted that the due-process clause does not apply when the government seizes property for law-enforcement purposes.[46] The Court did not buy this claim. In James Daniel Good, the Court reasoned that the government’s asserted interests in seizing property—obtaining jurisdiction and preventing the loss or destruction of the property—were slight in the case of real property. The Court ruled that due process requires the government to give notice and a hearing to owners of real property before seizures because of (1) the high risk of error created by allowing government to seize property without adhering to the requirements of due process of law, and (2) the substantial interests citizens have in being secure in their homes.\nThus, in just two terms of the Court, substantial constitutional constraints have been placed on government’s exercise of civil-forfeiture authority. Civil forfeiture remains, however, a source of revenue for government as it wages its war on drugs. For example, although James Daniel Good requires notice and a hearing prior to seizure of real property, the burden remains upon property owners to prove their innocence. Failure to prove innocence at a preseizure hearing means that government may seize the property and liquidate it for revenue purposes. We believe that more can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture. Indeed, as we read the Constitution, more must be done.\nAt a minimum, civil forfeitures of real property should all be declared unconstitutional. The common law of England and the United States at the time of the Constitution’s ratification did not provide for civil forfeiture of real properties; such forfeitures could be accomplished only following criminal convictions of the owners of real property. Unlike captured pirate ships and contraband goods, there is no chance that real property can stealthily escape jurisdiction before its owner is convicted. Consequently, the pragmatic rationale for allowing in rem seizures of personal property does not apply to real property. Because in rem forfeiture of real property was unheard of at the time the Constitution was ratified, the framers and ratifiers of the Constitution cannot plausibly be supposed to have intended to permit such forfeitures. Thus, any forfeitures of real property sought by government should be part of the explicit punishment sought in criminal proceedings.\nGiven the abuses that have marked civil forfeiture under the Drug Act, the Court might consider it wise to adopt an even more stringent approach to civil forfeiture. Under such an approach, the government would be barred from using civil procedures for the forfeiture of any property—real or personal—whenever (1) the property is forfeitable to the government because of its use in criminal wrongdoing, and (2) the government cannot prove that the property owner is beyond the personal jurisdiction of domestic courts. If the property owner is within the jurisdiction of domestic courts, then he or she should be duly accused, indicted, tried, and convicted before his or her property is forfeited to government. The protections of the criminal process are essential to safeguard innocent parties from government overreaching. If the owner is outside of domestic jurisdiction, the government must then, in good faith, attempt to notify the owner that his or her property temporarily (pending the outcome of a criminal trial) has been seized. If the owner then appears to wage a defense, a criminal conviction should be necessary before final forfeiture. In contrast, if the owner refuses to attend the trial, the property can then be forfeited in a civil action. In either case, however, the burden should be upon the government to prove wrongdoing by the owner, and not (as it is now) upon the owner to prove his or her innocence of any wrongdoing.\nGiven that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive in purpose?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture), the court should then ask, “Is the property contraband, or is it a legal good?” If it is contra-band, then the government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner.[47]\nWhen a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final. If the owner of the property is outside the court’s jurisdiction, then fair notice to the defendant of the charges is necessary, as is a fair opportunity for defense. If the defendant fails to show for his or her trial after being notified (or after a reasonable attempt to notify has been made), then a conviction obtained in absentia permits forfeiture to the government of the defendant’s property. The value of such forfeitures would still be limited, of course, by the excessive-fines clause of the Eighth Amendment.",
"role": "user"
},
{
"content": "The random manner in which civil-forfeiture\nare levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in\noffenses. This group is politically unorganized and\ngenerally cannot adequately defend itself against\nlobbies\nsuch as law-enforcement agencies\nwho clamor for\ncivil-forfeiture powers. Property owners\nhave little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes\nThe proceeds\nthus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement\nBecause the general population of taxpayers does not feel the full cost of government\nthe result is excessively\naggressive drug-law enforcement\nbiases result from exploitation of a revenue source that is\nhidden from general taxpayers\nAn alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement\nIn this scenario\nno one type of activity has a “revenue advantage” over others in attracting law-enforcement resources\nheads of law-enforcement agencies will thus make more rational enforcement decisions\nThe legislature is\nmost appropriate\nfor selecting the budget for law-enforcement agencies\nThe legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any\nspecialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or\nwelfare\nlegislatures are better placed to rationally allocate resources\nthan are individual enforcement agencies\nCivil forfeiture remains\nhowever, a source of revenue for government as it wages its war on drugs\nmore can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture\nmore must be done\nThe protections of the criminal process are essential to safeguard innocent parties from\noverreach\nGiven that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive\n?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture\nthe court should then ask, “Is the property contraband\n?” If it is\nthe government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner\nWhen a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing. | The random manner in which civil-forfeiture
are levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in
offenses. This group is politically unorganized and
generally cannot adequately defend itself against
lobbies
such as law-enforcement agencies
who clamor for
civil-forfeiture powers. Property owners
have little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes
The proceeds
thus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement
Because the general population of taxpayers does not feel the full cost of government
the result is excessively
aggressive drug-law enforcement
biases result from exploitation of a revenue source that is
hidden from general taxpayers
An alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement
In this scenario
no one type of activity has a “revenue advantage” over others in attracting law-enforcement resources
heads of law-enforcement agencies will thus make more rational enforcement decisions
The legislature is
most appropriate
for selecting the budget for law-enforcement agencies
The legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any
specialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or
welfare
legislatures are better placed to rationally allocate resources
than are individual enforcement agencies
Civil forfeiture remains
however, a source of revenue for government as it wages its war on drugs
more can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture
more must be done
The protections of the criminal process are essential to safeguard innocent parties from
overreach
Given that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive
?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture
the court should then ask, “Is the property contraband
?” If it is
the government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner
When a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final | random
civil-forfeiture
hides
taxpayers
part of the cost of government’s operations
suspect
politically unorganized
defend itself
lobbies
law-enforcement agencies
relieve
general body of taxpayers
collectively
public good of law enforcement
full cost
result
excessively
aggressive drug-law enforcement
biases
exploitation of a revenue source
hidden
general taxpayers
alternative
solely
legislatures
legislatures
entire
choose
quantity
no one type of activity has a “revenue advantage” over others
more rational enforcement decisions
legislature
most appropriate
large number
better positioned than any
specialized bureaucracies
schooling
welfare
better
rationally allocate resources
remains
revenue
more can be done by the courts
rein in
must
criminal process
essential
innocent parties
overreach
legislative legerdemain
a court should ask, “Is this forfeiture punitive
?”
no
contraband
?”
is
constitutionally seize
destroy
not contraband
absent a criminal conviction
taking
under the Fifth Amendment fair compensation must be paid to the dispossessed owner
is
all procedural requirements of the criminal law should be met before the forfeiture is final | 258 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing.",
"role": "system"
},
{
"content": "The random manner in which civil-forfeiture\nare levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in\noffenses. This group is politically unorganized and\ngenerally cannot adequately defend itself against\nlobbies\nsuch as law-enforcement agencies\nwho clamor for\ncivil-forfeiture powers. Property owners\nhave little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes\nThe proceeds\nthus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement\nBecause the general population of taxpayers does not feel the full cost of government\nthe result is excessively\naggressive drug-law enforcement\nbiases result from exploitation of a revenue source that is\nhidden from general taxpayers\nAn alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement\nIn this scenario\nno one type of activity has a “revenue advantage” over others in attracting law-enforcement resources\nheads of law-enforcement agencies will thus make more rational enforcement decisions\nThe legislature is\nmost appropriate\nfor selecting the budget for law-enforcement agencies\nThe legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any\nspecialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or\nwelfare\nlegislatures are better placed to rationally allocate resources\nthan are individual enforcement agencies\nCivil forfeiture remains\nhowever, a source of revenue for government as it wages its war on drugs\nmore can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture\nmore must be done\nThe protections of the criminal process are essential to safeguard innocent parties from\noverreach\nGiven that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive\n?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture\nthe court should then ask, “Is the property contraband\n?” If it is\nthe government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner\nWhen a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final",
"role": "user"
},
{
"content": "random\ncivil-forfeiture\nhides\ntaxpayers\npart of the cost of government’s operations\nsuspect\npolitically unorganized\ndefend itself\nlobbies\nlaw-enforcement agencies\nrelieve\ngeneral body of taxpayers\ncollectively\npublic good of law enforcement\nfull cost\nresult\nexcessively\naggressive drug-law enforcement\nbiases\nexploitation of a revenue source\nhidden\ngeneral taxpayers\nalternative\nsolely\nlegislatures\nlegislatures\nentire\nchoose\nquantity\nno one type of activity has a “revenue advantage” over others\nmore rational enforcement decisions\nlegislature\nmost appropriate\nlarge number\nbetter positioned than any\nspecialized bureaucracies\nschooling\nwelfare\nbetter\nrationally allocate resources\nremains\nrevenue\nmore can be done by the courts\nrein in\nmust\ncriminal process\nessential\ninnocent parties\noverreach\nlegislative legerdemain\na court should ask, “Is this forfeiture punitive\n?”\nno\ncontraband\n?”\nis\nconstitutionally seize\ndestroy\nnot contraband\nabsent a criminal conviction\ntaking\nunder the Fifth Amendment fair compensation must be paid to the dispossessed owner\nis\nall procedural requirements of the criminal law should be met before the forfeiture is final",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing.
Highlight Ratio: 0.2 | The random manner in which civil-forfeiture
are levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in
offenses. This group is politically unorganized and
generally cannot adequately defend itself against
lobbies
such as law-enforcement agencies
who clamor for
civil-forfeiture powers. Property owners
have little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes
The proceeds
thus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement
Because the general population of taxpayers does not feel the full cost of government
the result is excessively
aggressive drug-law enforcement
biases result from exploitation of a revenue source that is
hidden from general taxpayers
An alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement
In this scenario
no one type of activity has a “revenue advantage” over others in attracting law-enforcement resources
heads of law-enforcement agencies will thus make more rational enforcement decisions
The legislature is
most appropriate
for selecting the budget for law-enforcement agencies
The legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any
specialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or
welfare
legislatures are better placed to rationally allocate resources
than are individual enforcement agencies
Civil forfeiture remains
however, a source of revenue for government as it wages its war on drugs
more can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture
more must be done
The protections of the criminal process are essential to safeguard innocent parties from
overreach
Given that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive
?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture
the court should then ask, “Is the property contraband
?” If it is
the government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner
When a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final | civil-forfeiture
proceeds
relieve
general
taxpayers from
paying collectively for
law enforcement
the result is excessively
aggressive drug-law enforcement
An alternative
is to place budget decisions solely in
legislatures
In this
no one
has a “revenue advantage”
The legislature is
appropriate
confronts a large number of constituencies
better positioned than any
bureaucracies to determine if
revenue should be used
enforcement or
schooling or
welfare
rationally allocate resources
a court should ask, “Is this
punitive
If
no
ask, “Is the property contraband
If it is
the government can
seize
goods
if
not contraband, absent a criminal conviction, a taking
has occurred and
compensation must be paid | 259 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Third, it publicizes the costs of policing. Currently they are hidden because police are funding themselves through forfeitures. The CP forces them to ask legislatures for money to pay takings claims instead. Legislatures will refuse, reducing total forfeitures AND policing.\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "The random manner in which civil-forfeiture\nare levied hides from taxpayers a part of the cost of government’s operations. A disproportionate share of these costs are foisted upon owners whose properties law-enforcement agencies only suspect were used in\noffenses. This group is politically unorganized and\ngenerally cannot adequately defend itself against\nlobbies\nsuch as law-enforcement agencies\nwho clamor for\ncivil-forfeiture powers. Property owners\nhave little incentive to form or to join lobbying groups pressing for repeal or reform of civil-forfeiture statutes\nThe proceeds\nthus relieve the general body of taxpayers from some of the burden of paying collectively for the public good of law enforcement\nBecause the general population of taxpayers does not feel the full cost of government\nthe result is excessively\naggressive drug-law enforcement\nbiases result from exploitation of a revenue source that is\nhidden from general taxpayers\nAn alternative to giving government agencies discretion over their budgets is to place budget decisions solely in the hands of legislatures. When legislatures determine the entire budgets for law-enforcement agencies, legislatures in effect choose a quantity of law enforcement\nIn this scenario\nno one type of activity has a “revenue advantage” over others in attracting law-enforcement resources\nheads of law-enforcement agencies will thus make more rational enforcement decisions\nThe legislature is\nmost appropriate\nfor selecting the budget for law-enforcement agencies\nThe legislature confronts a large number of constituencies vying for public funds. Therefore, legislatures are better positioned than any\nspecialized bureaucracies to determine if an additional dollar of public revenue should be used for law enforcement or for schooling or\nwelfare\nlegislatures are better placed to rationally allocate resources\nthan are individual enforcement agencies\nCivil forfeiture remains\nhowever, a source of revenue for government as it wages its war on drugs\nmore can be done by the courts, under the Constitution, to further rein in government’s use of civil forfeiture\nmore must be done\nThe protections of the criminal process are essential to safeguard innocent parties from\noverreach\nGiven that one role of the courts is to protect citizens’ constitutional rights from legislative legerdemain, the label attached by the legislature to the forfeiture should be of no consequence. In all cases, a court should ask, “Is this forfeiture punitive\n?” If the answer is no (that is, if the court finds that no puni-tive purpose is served by the forfeiture\nthe court should then ask, “Is the property contraband\n?” If it is\nthe government can constitutionally seize these goods and destroy them. But if the property is not contraband, absent a criminal conviction, a taking of private property has occurred and under the Fifth Amendment fair compensation must be paid to the dispossessed owner\nWhen a court finds that the forfeiture is designed to be punitive, all procedural requirements of the criminal law should be met before the forfeiture is final",
"role": "user"
},
{
"content": "civil-forfeiture\nproceeds\nrelieve\ngeneral\ntaxpayers from\npaying collectively for\nlaw enforcement\nthe result is excessively\naggressive drug-law enforcement\nAn alternative\nis to place budget decisions solely in\nlegislatures\nIn this\nno one\nhas a “revenue advantage”\nThe legislature is\nappropriate\nconfronts a large number of constituencies\nbetter positioned than any\nbureaucracies to determine if\nrevenue should be used\nenforcement or\nschooling or\nwelfare\nrationally allocate resources\na court should ask, “Is this\npunitive\nIf\nno\nask, “Is the property contraband\nIf it is\nthe government can\nseize\ngoods\nif\nnot contraband, absent a criminal conviction, a taking\nhas occurred and\ncompensation must be paid",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: 2. LOWER COURTS---they’ll interpret the CP as an instruction. | August Mustillo 21, law student at the Campbell University School of Law and serves as a Staff Writer for the Campbell Law Observer, “Shedding Light on the Supreme Court’s Shadow Docket,” Campbell Law Observer, 4-2-2021, http://campbelllawobserver.com/shedding-light-on-the-supreme-courts-shadow-docket/
Furthermore, the Court needs flexibility for matters that arise unexpectedly, as in the example of COVID-19. Professor Michael Morley of Florida State University College of Law believes that utilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion that is likely to be treated as precedential. After all, a shadow docket ruling will likely not have the same weight as a full written opinion. Instead, some lower courts view these orders more as “signals.” | the Court needs flexibility for matters that arise unexpectedly
utilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion
a shadow docket ruling
lower courts view
as “signals.” | 260 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: 2. LOWER COURTS---they’ll interpret the CP as an instruction.",
"role": "system"
},
{
"content": "August Mustillo 21, law student at the Campbell University School of Law and serves as a Staff Writer for the Campbell Law Observer, “Shedding Light on the Supreme Court’s Shadow Docket,” Campbell Law Observer, 4-2-2021, http://campbelllawobserver.com/shedding-light-on-the-supreme-courts-shadow-docket/\nFurthermore, the Court needs flexibility for matters that arise unexpectedly, as in the example of COVID-19. Professor Michael Morley of Florida State University College of Law believes that utilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion that is likely to be treated as precedential. After all, a shadow docket ruling will likely not have the same weight as a full written opinion. Instead, some lower courts view these orders more as “signals.”",
"role": "user"
},
{
"content": "the Court needs flexibility for matters that arise unexpectedly\nutilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion\na shadow docket ruling\nlower courts view\nas “signals.”",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: 2. LOWER COURTS---they’ll interpret the CP as an instruction. | the Court needs flexibility for matters that arise unexpectedly
utilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion
a shadow docket ruling
lower courts view
as “signals.” | flex
unexpectedly
getting the Justices to agree on an ultimate outcome
more beneficial
hurried opinion
ruling
lower courts
“signals.” | 261 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: 2. LOWER COURTS---they’ll interpret the CP as an instruction.",
"role": "system"
},
{
"content": "the Court needs flexibility for matters that arise unexpectedly\nutilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion\na shadow docket ruling\nlower courts view\nas “signals.”",
"role": "user"
},
{
"content": "flex\nunexpectedly\ngetting the Justices to agree on an ultimate outcome\nmore beneficial\nhurried opinion\nruling\nlower courts\n“signals.”",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: 2. LOWER COURTS---they’ll interpret the CP as an instruction.
Highlight Ratio: 0.6 | the Court needs flexibility for matters that arise unexpectedly
utilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion
a shadow docket ruling
lower courts view
as “signals.” | utilizing the
shadow docket for
getting
Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion
a shadow docket ruling
lower courts view
as “signals.” | 262 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: 2. LOWER COURTS---they’ll interpret the CP as an instruction.\n\nHighlight Ratio: 0.6",
"role": "system"
},
{
"content": "the Court needs flexibility for matters that arise unexpectedly\nutilizing the nature of the shadow docket for purposes of simply getting the Justices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion\na shadow docket ruling\nlower courts view\nas “signals.”",
"role": "user"
},
{
"content": "utilizing the\nshadow docket for\ngetting\nJustices to agree on an ultimate outcome may be more beneficial than issuing a hurried opinion\na shadow docket ruling\nlower courts view\nas “signals.”",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: That means the CP will be treated as binding. | When issuing “signals,” the Justices act in their official, adjudicatory capacities without establishing conventional precedent, but nonetheless indicate some aspect of how lower courts should decide cases.101 For example, signals can arise in summary orders, statements during oral argument, separate opinions, and dicta in majority opinions. Signals are thus distinct from the many informal cues that arise outside of the traditional adjudicatory process, such as individual Justices’ statements to the press.102
The signals model maintains that signals from a majority of the Justices have precedential force but are subordinate to conventional precedent.103 [FOOTNOTE 103 BEGINS]103. Scholars have documented the Court’s use of signals to influence its own docket, particularly by informing sophisticated nonjudicial actors of what type of cases the Justices want to decide. See VANESSA A. BAIRD, ANSWERING THE CALL OF THE COURT: HOW JUSTICES AND LITIGANTS SET THE SUPREME COURT AGENDA (2007); Tonja Jacobi, The Judicial Signaling Game: How Judges Shape Their Dockets, 16 SUP. CT. ECON. REV. 1, 2 (2008) (discussing the idea that “judges signal the outcome of future cases in order to actively shape their dockets”). In these accounts, lower courts play a secondary or facilitative role. See, e.g., BAIRD, supra, at 70 (“[S]ignals from the Supreme Court create an incentive for appeals courts to provide information that can help the Supreme Court justices discern which cases are likely to be the best vehicles for political or legal change.”); H. W. PERRY, JR., DECIDING TO DECIDE: AGENDA SETTING IN THE UNITED STATES SUPREME COURT 177 (1991). [FOOTNOTE 103 ENDS]This two-tiered approach to vertical stare decisis builds on the authority model in treating the Court’s official efforts to resolve cases and controversies—and nothing else—as authoritative. In this respect, the signals model diverges from the prediction model’s forward-looking orientation and openness to considering the Justices’ extrajudicial statements as grounds for prediction. At the same time, the signals model captures the main pragmatic benefits of the predictive and proficiency models by attributing supplemental precedential force to the Court’s relatively informal actions, which are likely to be less proficient and predictive than conventional precedent. This approach allows the Court to honor its traditional adjudicatory role while more effectively managing lower courts.
Although the Court’s habit of sending nonprecedential signals is a relatively unremarked-on phenomenon,104 the dearth of commentary on this point largely reflects changing times. Because it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible today largely because of contemporary digital media and the growing culture of Court-watching.105 Before the Internet, it wasn’t really possible for far-flung courts and advocates to keep track of the Court’s stay orders, oral arguments, and the like. Indeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports.106 In the Court’s early years, most local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management.107 Once circuit riding ended, the Court’s formally published precedents became its primary mode of vertical control. Today, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media such as blogs and Twitter. And the rise of the Supreme Court bar, which counts former clerks among its members, has fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts while still acting like an adjudicatory body. With their levers of influence multiplied, the Justices might naturally explore new opportunities to use them.108
In using signals to exert managerial oversight outside the conventional bounds of precedent, the Court often promotes efficiency and uniformity. In many cases, the Justices implicitly decide or expressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions or other preliminary rulings made below; or they may pertain to substantive legal questions not presently before the Court.109 In other cases, the Court comments negatively on a disfavored precedent, such as by asserting that the precedent is “narrow” or that it is difficult to prevail under the precedent.110 Relatedly, the Court sometimes establishes a pattern of repeatedly narrowing a precedent, thereby tacitly establishing the precedent’s disfavored status.111 Because these decisions, comments, and patterns lie outside the bounds of conventional precedent, they are often treated as dicta or otherwise denied precedential status. Nonetheless, the Justices routinely express deliberate views on ancillary topics in separate opinions or during oral argument, with the apparent intention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”112 [FOOTNOTE 112 BEGINS]112. For a recent example, the Eighth Circuit expressly followed the signal that the Court had sent in stay orders relating to the religious exemption from the Affordable Care Act’s contraception coverage provisions. See Sharpe Holdings, Inc. v. U.S. Dep’t of Health & Human Servs., 801 F.3d 927, 944 (8th Cir. 2015) (“Although the Court’s orders were not final rulings on the merits, they at the very least collectively constitute a signal that less restrictive means exist by which the government may further its interests.” (citing Priests for Life v. U.S. Dep’t of Health & Human Servs., 808 F.3d 1, 25 (D.C. Cir. 2015) (Kavanaugh, J., dissenting from denial of rehearing en banc)); see also Priests for Life v. U.S. Dep’t of Health & Human Servs., 772 F.3d 229, 257 (D.C. Cir. 2014) (finding no substantial burden but acknowledging “that the Supreme Court’s recent order in Wheaton College might be read to signal a different conclusion”); DeBoer v. Snyder, 772 F.3d 388, 402 (6th Cir. 2014) (“But don’t these denials of certiorari signal that, from the Court’s perspective, the right to same-sex marriage is inevitable? Maybe; maybe not.”). For examples of signals during oral argument, see Re, supra note 101. [FOOTNOTE 112 ENDS]
The main difficulty with the signals model is the challenge of reliably identifying signals. This problem arises precisely because signaling is marked by the lack of formality associated with full-dress rulings on the merits. That the Court is so closely observed is part of what makes signaling possible, but that reality also multiplies the opportunities for picking up false signals. And because so many people have an interest in gleaning desired signals from the Court, there is a serious risk that self-interested observers will misinterpret signals.113 Complicating things even further, Justices on the losing end of cases might sometimes send false signals by suggesting that they speak for a majority while actually offering just noise or cheap talk that is better ignored than relied on.
Yet lower courts can mitigate signals’ downsides. First, signals should give way in the face of clear conventional precedent. This principle reflects the comparative rigor, deliberateness, and clarity associated with conventional holdings and limits the opportunity for mistaken signals. In other words, signals cannot justify lower courts in overruling or partially overruling higher court precedent. But when conventional higher court precedent is ambiguous, lower courts should follow signals to resolve that ambiguity, including by narrowing the conventional precedent from below. In this way, the signals model depends on the recognition that narrowing from below is distinct from anticipatory overruling. Second, lower courts should attend to signals only when they reflect deliberate judgments by a majority of the Justices. This restriction addresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole. | When issuing “signals,”
Justices act in their official
capacities without establishing conventional precedent, but nonetheless indicate
how lower courts should decide
signals can arise in summary orders
are thus distinct from
informal cues that arise outside of the traditional adjudicatory process, such as
Justices’ statements to the press
signals from a majority of the Justices have precedential force
Scholars have documented the Court’s use of signals to influence its own docket, particularly by informing
nonjudicial actors
what type of cases the Justices want to decide
judges signal the outcome of future cases in order to actively shape
dockets
lower courts play a
facilitative role
[S]ignals
create an incentive for appeals courts to provide information that
help the
Court
discern which
are likely
the best vehicles for
change
This two-tiered approach
builds on the authority model in treating
official efforts to resolve cases
as authoritative
attributing
precedential force to the Court’s relatively informal actions
Because it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible
because of contemporary digital media and the growing culture of Court-watching
Before the Internet, it wasn’t really possible for far-flung courts and advocates to keep track
Indeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports
most local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management
Today, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media
the Supreme Court bar
has fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts
In using signals
the Court
promotes efficiency and uniformity
the
expressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions
Because these
patterns lie outside the bounds of conventional precedent, they are often treated as dicta
Nonetheless, the Justices routinely express deliberate views on ancillary topics
with the
intention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”
For a recent example, the Eighth Circuit expressly followed the signal
the Court
sent in stay orders
Although the Court’s orders were not final rulings on the merits, they
collectively constitute a signal that less restrictive means exist by which the government may further its interests
lower courts can mitigate signals’ downsides
when
precedent is ambiguous, lower courts should follow signals to resolve
ambiguity
by narrowing
conventional precedent from below
lower courts
attend to signals only when they reflect deliberate judgments by a majority
This
addresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole. | 263 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: That means the CP will be treated as binding.",
"role": "system"
},
{
"content": "When issuing “signals,” the Justices act in their official, adjudicatory capacities without establishing conventional precedent, but nonetheless indicate some aspect of how lower courts should decide cases.101 For example, signals can arise in summary orders, statements during oral argument, separate opinions, and dicta in majority opinions. Signals are thus distinct from the many informal cues that arise outside of the traditional adjudicatory process, such as individual Justices’ statements to the press.102\nThe signals model maintains that signals from a majority of the Justices have precedential force but are subordinate to conventional precedent.103 [FOOTNOTE 103 BEGINS]103. Scholars have documented the Court’s use of signals to influence its own docket, particularly by informing sophisticated nonjudicial actors of what type of cases the Justices want to decide. See VANESSA A. BAIRD, ANSWERING THE CALL OF THE COURT: HOW JUSTICES AND LITIGANTS SET THE SUPREME COURT AGENDA (2007); Tonja Jacobi, The Judicial Signaling Game: How Judges Shape Their Dockets, 16 SUP. CT. ECON. REV. 1, 2 (2008) (discussing the idea that “judges signal the outcome of future cases in order to actively shape their dockets”). In these accounts, lower courts play a secondary or facilitative role. See, e.g., BAIRD, supra, at 70 (“[S]ignals from the Supreme Court create an incentive for appeals courts to provide information that can help the Supreme Court justices discern which cases are likely to be the best vehicles for political or legal change.”); H. W. PERRY, JR., DECIDING TO DECIDE: AGENDA SETTING IN THE UNITED STATES SUPREME COURT 177 (1991). [FOOTNOTE 103 ENDS]This two-tiered approach to vertical stare decisis builds on the authority model in treating the Court’s official efforts to resolve cases and controversies—and nothing else—as authoritative. In this respect, the signals model diverges from the prediction model’s forward-looking orientation and openness to considering the Justices’ extrajudicial statements as grounds for prediction. At the same time, the signals model captures the main pragmatic benefits of the predictive and proficiency models by attributing supplemental precedential force to the Court’s relatively informal actions, which are likely to be less proficient and predictive than conventional precedent. This approach allows the Court to honor its traditional adjudicatory role while more effectively managing lower courts.\nAlthough the Court’s habit of sending nonprecedential signals is a relatively unremarked-on phenomenon,104 the dearth of commentary on this point largely reflects changing times. Because it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible today largely because of contemporary digital media and the growing culture of Court-watching.105 Before the Internet, it wasn’t really possible for far-flung courts and advocates to keep track of the Court’s stay orders, oral arguments, and the like. Indeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports.106 In the Court’s early years, most local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management.107 Once circuit riding ended, the Court’s formally published precedents became its primary mode of vertical control. Today, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media such as blogs and Twitter. And the rise of the Supreme Court bar, which counts former clerks among its members, has fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts while still acting like an adjudicatory body. With their levers of influence multiplied, the Justices might naturally explore new opportunities to use them.108\nIn using signals to exert managerial oversight outside the conventional bounds of precedent, the Court often promotes efficiency and uniformity. In many cases, the Justices implicitly decide or expressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions or other preliminary rulings made below; or they may pertain to substantive legal questions not presently before the Court.109 In other cases, the Court comments negatively on a disfavored precedent, such as by asserting that the precedent is “narrow” or that it is difficult to prevail under the precedent.110 Relatedly, the Court sometimes establishes a pattern of repeatedly narrowing a precedent, thereby tacitly establishing the precedent’s disfavored status.111 Because these decisions, comments, and patterns lie outside the bounds of conventional precedent, they are often treated as dicta or otherwise denied precedential status. Nonetheless, the Justices routinely express deliberate views on ancillary topics in separate opinions or during oral argument, with the apparent intention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”112 [FOOTNOTE 112 BEGINS]112. For a recent example, the Eighth Circuit expressly followed the signal that the Court had sent in stay orders relating to the religious exemption from the Affordable Care Act’s contraception coverage provisions. See Sharpe Holdings, Inc. v. U.S. Dep’t of Health & Human Servs., 801 F.3d 927, 944 (8th Cir. 2015) (“Although the Court’s orders were not final rulings on the merits, they at the very least collectively constitute a signal that less restrictive means exist by which the government may further its interests.” (citing Priests for Life v. U.S. Dep’t of Health & Human Servs., 808 F.3d 1, 25 (D.C. Cir. 2015) (Kavanaugh, J., dissenting from denial of rehearing en banc)); see also Priests for Life v. U.S. Dep’t of Health & Human Servs., 772 F.3d 229, 257 (D.C. Cir. 2014) (finding no substantial burden but acknowledging “that the Supreme Court’s recent order in Wheaton College might be read to signal a different conclusion”); DeBoer v. Snyder, 772 F.3d 388, 402 (6th Cir. 2014) (“But don’t these denials of certiorari signal that, from the Court’s perspective, the right to same-sex marriage is inevitable? Maybe; maybe not.”). For examples of signals during oral argument, see Re, supra note 101. [FOOTNOTE 112 ENDS]\nThe main difficulty with the signals model is the challenge of reliably identifying signals. This problem arises precisely because signaling is marked by the lack of formality associated with full-dress rulings on the merits. That the Court is so closely observed is part of what makes signaling possible, but that reality also multiplies the opportunities for picking up false signals. And because so many people have an interest in gleaning desired signals from the Court, there is a serious risk that self-interested observers will misinterpret signals.113 Complicating things even further, Justices on the losing end of cases might sometimes send false signals by suggesting that they speak for a majority while actually offering just noise or cheap talk that is better ignored than relied on.\nYet lower courts can mitigate signals’ downsides. First, signals should give way in the face of clear conventional precedent. This principle reflects the comparative rigor, deliberateness, and clarity associated with conventional holdings and limits the opportunity for mistaken signals. In other words, signals cannot justify lower courts in overruling or partially overruling higher court precedent. But when conventional higher court precedent is ambiguous, lower courts should follow signals to resolve that ambiguity, including by narrowing the conventional precedent from below. In this way, the signals model depends on the recognition that narrowing from below is distinct from anticipatory overruling. Second, lower courts should attend to signals only when they reflect deliberate judgments by a majority of the Justices. This restriction addresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole.",
"role": "user"
},
{
"content": "When issuing “signals,”\nJustices act in their official\ncapacities without establishing conventional precedent, but nonetheless indicate\nhow lower courts should decide\nsignals can arise in summary orders\nare thus distinct from\ninformal cues that arise outside of the traditional adjudicatory process, such as\nJustices’ statements to the press\nsignals from a majority of the Justices have precedential force\nScholars have documented the Court’s use of signals to influence its own docket, particularly by informing\nnonjudicial actors\nwhat type of cases the Justices want to decide\njudges signal the outcome of future cases in order to actively shape\ndockets\nlower courts play a\nfacilitative role\n[S]ignals\ncreate an incentive for appeals courts to provide information that\nhelp the\nCourt\ndiscern which\nare likely\nthe best vehicles for\nchange\nThis two-tiered approach\nbuilds on the authority model in treating\nofficial efforts to resolve cases\nas authoritative\nattributing\nprecedential force to the Court’s relatively informal actions\nBecause it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible\nbecause of contemporary digital media and the growing culture of Court-watching\nBefore the Internet, it wasn’t really possible for far-flung courts and advocates to keep track\nIndeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports\nmost local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management\nToday, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media\nthe Supreme Court bar\nhas fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts\nIn using signals\nthe Court\npromotes efficiency and uniformity\nthe\nexpressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions\nBecause these\npatterns lie outside the bounds of conventional precedent, they are often treated as dicta\nNonetheless, the Justices routinely express deliberate views on ancillary topics\nwith the\nintention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”\nFor a recent example, the Eighth Circuit expressly followed the signal\nthe Court\nsent in stay orders\nAlthough the Court’s orders were not final rulings on the merits, they\ncollectively constitute a signal that less restrictive means exist by which the government may further its interests\nlower courts can mitigate signals’ downsides\nwhen\nprecedent is ambiguous, lower courts should follow signals to resolve\nambiguity\nby narrowing\nconventional precedent from below\nlower courts\nattend to signals only when they reflect deliberate judgments by a majority\nThis\naddresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: That means the CP will be treated as binding. | When issuing “signals,”
Justices act in their official
capacities without establishing conventional precedent, but nonetheless indicate
how lower courts should decide
signals can arise in summary orders
are thus distinct from
informal cues that arise outside of the traditional adjudicatory process, such as
Justices’ statements to the press
signals from a majority of the Justices have precedential force
Scholars have documented the Court’s use of signals to influence its own docket, particularly by informing
nonjudicial actors
what type of cases the Justices want to decide
judges signal the outcome of future cases in order to actively shape
dockets
lower courts play a
facilitative role
[S]ignals
create an incentive for appeals courts to provide information that
help the
Court
discern which
are likely
the best vehicles for
change
This two-tiered approach
builds on the authority model in treating
official efforts to resolve cases
as authoritative
attributing
precedential force to the Court’s relatively informal actions
Because it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible
because of contemporary digital media and the growing culture of Court-watching
Before the Internet, it wasn’t really possible for far-flung courts and advocates to keep track
Indeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports
most local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management
Today, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media
the Supreme Court bar
has fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts
In using signals
the Court
promotes efficiency and uniformity
the
expressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions
Because these
patterns lie outside the bounds of conventional precedent, they are often treated as dicta
Nonetheless, the Justices routinely express deliberate views on ancillary topics
with the
intention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”
For a recent example, the Eighth Circuit expressly followed the signal
the Court
sent in stay orders
Although the Court’s orders were not final rulings on the merits, they
collectively constitute a signal that less restrictive means exist by which the government may further its interests
lower courts can mitigate signals’ downsides
when
precedent is ambiguous, lower courts should follow signals to resolve
ambiguity
by narrowing
conventional precedent from below
lower courts
attend to signals only when they reflect deliberate judgments by a majority
This
addresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole. | “signals,”
official
capacities
conventional
nonetheless indicate
how lower courts should decide
summary orders
distinct
informal
outside
majority of the Justices
precedential force
documented
own docket
informing
nonjudicial actors
what type of cases the Justices want to decide
signal
future cases
actively shape
dockets
facilitative role
incentive
provide information
vehicles for
change
two-tiered approach
authority
resolve
authoritative
precedential force
relatively
peripheral
summary decisions
discretionary relief
feasible
digital media
Court-watching
Before
track
printed
limited information
best means
lots
follow the Court
digital media
Supreme Court bar
intense attention
greater opportunity
guidance
signals
efficiency
uniformity
expressly opine
ancillary issues
procedure
stay
outside
treated
dicta
deliberate views
intention
lower courts
pick up the message
often do just that
even using the term “signal.”
recent example
expressly followed the signal
the Court
sent in stay orders
not final rulings on the merits
collectively
signal
less restrictive means
lower courts
mitigate signals’ downsides
ambiguous
follow signals to resolve
ambiguity
narrowing
conventional
below
only
deliberate judgments
majority | 264 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: That means the CP will be treated as binding.",
"role": "system"
},
{
"content": "When issuing “signals,”\nJustices act in their official\ncapacities without establishing conventional precedent, but nonetheless indicate\nhow lower courts should decide\nsignals can arise in summary orders\nare thus distinct from\ninformal cues that arise outside of the traditional adjudicatory process, such as\nJustices’ statements to the press\nsignals from a majority of the Justices have precedential force\nScholars have documented the Court’s use of signals to influence its own docket, particularly by informing\nnonjudicial actors\nwhat type of cases the Justices want to decide\njudges signal the outcome of future cases in order to actively shape\ndockets\nlower courts play a\nfacilitative role\n[S]ignals\ncreate an incentive for appeals courts to provide information that\nhelp the\nCourt\ndiscern which\nare likely\nthe best vehicles for\nchange\nThis two-tiered approach\nbuilds on the authority model in treating\nofficial efforts to resolve cases\nas authoritative\nattributing\nprecedential force to the Court’s relatively informal actions\nBecause it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible\nbecause of contemporary digital media and the growing culture of Court-watching\nBefore the Internet, it wasn’t really possible for far-flung courts and advocates to keep track\nIndeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports\nmost local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management\nToday, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media\nthe Supreme Court bar\nhas fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts\nIn using signals\nthe Court\npromotes efficiency and uniformity\nthe\nexpressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions\nBecause these\npatterns lie outside the bounds of conventional precedent, they are often treated as dicta\nNonetheless, the Justices routinely express deliberate views on ancillary topics\nwith the\nintention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”\nFor a recent example, the Eighth Circuit expressly followed the signal\nthe Court\nsent in stay orders\nAlthough the Court’s orders were not final rulings on the merits, they\ncollectively constitute a signal that less restrictive means exist by which the government may further its interests\nlower courts can mitigate signals’ downsides\nwhen\nprecedent is ambiguous, lower courts should follow signals to resolve\nambiguity\nby narrowing\nconventional precedent from below\nlower courts\nattend to signals only when they reflect deliberate judgments by a majority\nThis\naddresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole.",
"role": "user"
},
{
"content": "“signals,”\nofficial\ncapacities\nconventional\nnonetheless indicate\nhow lower courts should decide\nsummary orders\ndistinct\ninformal\noutside\nmajority of the Justices\nprecedential force\ndocumented\nown docket\ninforming\nnonjudicial actors\nwhat type of cases the Justices want to decide\nsignal\nfuture cases\nactively shape\ndockets\nfacilitative role\nincentive\nprovide information\nvehicles for\nchange\ntwo-tiered approach\nauthority\nresolve\nauthoritative\nprecedential force\nrelatively\nperipheral\nsummary decisions\ndiscretionary relief\nfeasible\ndigital media\nCourt-watching\nBefore\ntrack\nprinted\nlimited information\nbest means\nlots\nfollow the Court\ndigital media\nSupreme Court bar\nintense attention\ngreater opportunity\nguidance\nsignals\nefficiency\nuniformity\nexpressly opine\nancillary issues\nprocedure\nstay\noutside\ntreated\ndicta\ndeliberate views\nintention\nlower courts\npick up the message\noften do just that\neven using the term “signal.”\nrecent example\nexpressly followed the signal\nthe Court\nsent in stay orders\nnot final rulings on the merits\ncollectively\nsignal\nless restrictive means\nlower courts\nmitigate signals’ downsides\nambiguous\nfollow signals to resolve\nambiguity\nnarrowing\nconventional\nbelow\nonly\ndeliberate judgments\nmajority",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: That means the CP will be treated as binding.
Highlight Ratio: 0.3 | When issuing “signals,”
Justices act in their official
capacities without establishing conventional precedent, but nonetheless indicate
how lower courts should decide
signals can arise in summary orders
are thus distinct from
informal cues that arise outside of the traditional adjudicatory process, such as
Justices’ statements to the press
signals from a majority of the Justices have precedential force
Scholars have documented the Court’s use of signals to influence its own docket, particularly by informing
nonjudicial actors
what type of cases the Justices want to decide
judges signal the outcome of future cases in order to actively shape
dockets
lower courts play a
facilitative role
[S]ignals
create an incentive for appeals courts to provide information that
help the
Court
discern which
are likely
the best vehicles for
change
This two-tiered approach
builds on the authority model in treating
official efforts to resolve cases
as authoritative
attributing
precedential force to the Court’s relatively informal actions
Because it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible
because of contemporary digital media and the growing culture of Court-watching
Before the Internet, it wasn’t really possible for far-flung courts and advocates to keep track
Indeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports
most local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management
Today, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media
the Supreme Court bar
has fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts
In using signals
the Court
promotes efficiency and uniformity
the
expressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions
Because these
patterns lie outside the bounds of conventional precedent, they are often treated as dicta
Nonetheless, the Justices routinely express deliberate views on ancillary topics
with the
intention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”
For a recent example, the Eighth Circuit expressly followed the signal
the Court
sent in stay orders
Although the Court’s orders were not final rulings on the merits, they
collectively constitute a signal that less restrictive means exist by which the government may further its interests
lower courts can mitigate signals’ downsides
when
precedent is ambiguous, lower courts should follow signals to resolve
ambiguity
by narrowing
conventional precedent from below
lower courts
attend to signals only when they reflect deliberate judgments by a majority
This
addresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole. | When issuing “signals,”
Justices act in
official
capacities without establishing conventional precedent, but
indicate
how
courts should decide
can arise in summary orders
attributing
precedential force to
relatively informal actions
Because it can occur through peripheral actions
signaling is feasible
because of
digital media and
Court-watching
lots of people, including advocates and
judges, follow the Court
the
bar
fostered intense attention to
details of
operations. The result is
greater opportunity to provide guidance
the Court
expressly opine on ancillary issues
they are often treated as dicta
Nonetheless
lower courts
pick up the message
even using the term “signal.”
the Eighth Circuit expressly followed the signal
in stay orders
when
precedent is ambiguous, lower courts
follow signals to resolve
ambiguity
by narrowing
precedent from below | 265 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: That means the CP will be treated as binding.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "When issuing “signals,”\nJustices act in their official\ncapacities without establishing conventional precedent, but nonetheless indicate\nhow lower courts should decide\nsignals can arise in summary orders\nare thus distinct from\ninformal cues that arise outside of the traditional adjudicatory process, such as\nJustices’ statements to the press\nsignals from a majority of the Justices have precedential force\nScholars have documented the Court’s use of signals to influence its own docket, particularly by informing\nnonjudicial actors\nwhat type of cases the Justices want to decide\njudges signal the outcome of future cases in order to actively shape\ndockets\nlower courts play a\nfacilitative role\n[S]ignals\ncreate an incentive for appeals courts to provide information that\nhelp the\nCourt\ndiscern which\nare likely\nthe best vehicles for\nchange\nThis two-tiered approach\nbuilds on the authority model in treating\nofficial efforts to resolve cases\nas authoritative\nattributing\nprecedential force to the Court’s relatively informal actions\nBecause it can occur through peripheral actions like summary decisions on discretionary relief, signaling is feasible\nbecause of contemporary digital media and the growing culture of Court-watching\nBefore the Internet, it wasn’t really possible for far-flung courts and advocates to keep track\nIndeed, for much of U.S. history it was hard enough for the Court to exert managerial influence through printed case reports\nmost local judges had at best limited information on the Court’s decisions, and the Justices’ practice of circuit riding offered perhaps the best means of informal judicial management\nToday, however, lots of people, including advocates and lower court judges, follow the Court with the help of digital media\nthe Supreme Court bar\nhas fostered intense attention to the details of the Court’s operations. The result is that the Court has greater opportunity to provide guidance to lower courts\nIn using signals\nthe Court\npromotes efficiency and uniformity\nthe\nexpressly opine on ancillary issues while resolving the case at hand. These decisions may concern matters of procedure, such as stay decisions\nBecause these\npatterns lie outside the bounds of conventional precedent, they are often treated as dicta\nNonetheless, the Justices routinely express deliberate views on ancillary topics\nwith the\nintention that lower courts will pick up the message. And the lower courts often do just that, sometimes even using the term “signal.”\nFor a recent example, the Eighth Circuit expressly followed the signal\nthe Court\nsent in stay orders\nAlthough the Court’s orders were not final rulings on the merits, they\ncollectively constitute a signal that less restrictive means exist by which the government may further its interests\nlower courts can mitigate signals’ downsides\nwhen\nprecedent is ambiguous, lower courts should follow signals to resolve\nambiguity\nby narrowing\nconventional precedent from below\nlower courts\nattend to signals only when they reflect deliberate judgments by a majority\nThis\naddresses the risk that individual Justices might attempt to influence lower courts by telegraphing minority views that may even be opposed by the Court as a whole.",
"role": "user"
},
{
"content": "When issuing “signals,”\nJustices act in\nofficial\ncapacities without establishing conventional precedent, but\nindicate\nhow\ncourts should decide\ncan arise in summary orders\nattributing\nprecedential force to\nrelatively informal actions\nBecause it can occur through peripheral actions\nsignaling is feasible\nbecause of\ndigital media and\nCourt-watching\nlots of people, including advocates and\njudges, follow the Court\nthe\nbar\nfostered intense attention to\ndetails of\noperations. The result is\ngreater opportunity to provide guidance\nthe Court\nexpressly opine on ancillary issues\nthey are often treated as dicta\nNonetheless\nlower courts\npick up the message\neven using the term “signal.”\nthe Eighth Circuit expressly followed the signal\nin stay orders\nwhen\nprecedent is ambiguous, lower courts\nfollow signals to resolve\nambiguity\nby narrowing\nprecedent from below",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Recent streamlining is improving resource allocation | The Federal Trade Commission voted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate seven specific enforcement priorities. Priority targets include repeat offenders; technology companies and digital platforms; and healthcare businesses such as pharmaceutical companies, pharmacy benefits managers, and hospitals. The agency is also prioritizing investigations into harms against workers and small businesses, along with harms related to the COVID-19 pandemic. Finally, at a time when merger filings are surging, the agency is ramping up enforcement against illegal mergers, both proposed and consummated.
In remarks delivered during the open meeting, Chair Lina M. Khan noted that the resolutions approved today represent an important step in rethinking the work of the FTC. Instituting new cross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors across markets.
“The reforms are designed to ensure that our staff can comprehensively investigate unlawful business practices across the economy,” said Chair Khan. “They will help relieve unnecessary burdens on staff and cut back delays and ‘red tape’ bureaucracy when it comes to advancing our Commission’s law enforcement priorities. This is particularly important given that we are in the midst of a massive merger boom.”
Compulsory process refers to the issuance of demands for documents and testimony, through the use of civil investigative demands and subpoena. The FTC Act authorizes the Commission to use compulsory process in its investigations. Compulsory process requires the recipient to produce information, and these orders are enforceable by courts. The Commission has routinely adopted compulsory process resolutions on a wide range of topics. Many of these resolutions cover specific industries, like the automobile industry or the postsecondary education industry, while others involve business practices that cut across sectors, like privacy or the targeting of older Americans.
The actions taken today will broaden the ability for FTC investigators and prosecutors to obtain evidence in critical investigations on key areas where the FTC’s work can make the most impact. Each omnibus authorizes investigations into any competition or consumer protection conduct violations under the FTC Act. The omnibuses will also allow staff to use compulsory process to investigate both proposed mergers and consummated mergers. Individual Commissioners will continue to be required to sign compulsory process documents prior to issuance. With these in place, the FTC can better utilize its limited resources and move forward in earnest to fix the market structures that allow the worst predators to proliferate. | The F
T
C
voted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate
enforcement priorities
the resolutions
represent an important step in rethinking the work of the FTC
cross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors
The actions taken
will
allow staff to use compulsory process to investigate both proposed mergers and consummated mergers
With these in place, the FTC can better utilize its limited resources and move forward in earnest to fix
market structures | 266 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Recent streamlining is improving resource allocation",
"role": "system"
},
{
"content": "The Federal Trade Commission voted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate seven specific enforcement priorities. Priority targets include repeat offenders; technology companies and digital platforms; and healthcare businesses such as pharmaceutical companies, pharmacy benefits managers, and hospitals. The agency is also prioritizing investigations into harms against workers and small businesses, along with harms related to the COVID-19 pandemic. Finally, at a time when merger filings are surging, the agency is ramping up enforcement against illegal mergers, both proposed and consummated.\nIn remarks delivered during the open meeting, Chair Lina M. Khan noted that the resolutions approved today represent an important step in rethinking the work of the FTC. Instituting new cross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors across markets.\n“The reforms are designed to ensure that our staff can comprehensively investigate unlawful business practices across the economy,” said Chair Khan. “They will help relieve unnecessary burdens on staff and cut back delays and ‘red tape’ bureaucracy when it comes to advancing our Commission’s law enforcement priorities. This is particularly important given that we are in the midst of a massive merger boom.”\nCompulsory process refers to the issuance of demands for documents and testimony, through the use of civil investigative demands and subpoena. The FTC Act authorizes the Commission to use compulsory process in its investigations. Compulsory process requires the recipient to produce information, and these orders are enforceable by courts. The Commission has routinely adopted compulsory process resolutions on a wide range of topics. Many of these resolutions cover specific industries, like the automobile industry or the postsecondary education industry, while others involve business practices that cut across sectors, like privacy or the targeting of older Americans.\nThe actions taken today will broaden the ability for FTC investigators and prosecutors to obtain evidence in critical investigations on key areas where the FTC’s work can make the most impact. Each omnibus authorizes investigations into any competition or consumer protection conduct violations under the FTC Act. The omnibuses will also allow staff to use compulsory process to investigate both proposed mergers and consummated mergers. Individual Commissioners will continue to be required to sign compulsory process documents prior to issuance. With these in place, the FTC can better utilize its limited resources and move forward in earnest to fix the market structures that allow the worst predators to proliferate.",
"role": "user"
},
{
"content": "The F\nT\nC\nvoted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate\nenforcement priorities\nthe resolutions\nrepresent an important step in rethinking the work of the FTC\ncross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors\nThe actions taken\nwill\nallow staff to use compulsory process to investigate both proposed mergers and consummated mergers\nWith these in place, the FTC can better utilize its limited resources and move forward in earnest to fix\nmarket structures",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Recent streamlining is improving resource allocation
Highlight Ratio: 0.3 | The F
T
C
voted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate
enforcement priorities
the resolutions
represent an important step in rethinking the work of the FTC
cross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors
The actions taken
will
allow staff to use compulsory process to investigate both proposed mergers and consummated mergers
With these in place, the FTC can better utilize its limited resources and move forward in earnest to fix
market structures | The F
T
C
approve
resolutions authorizing
key
enforcement priorities
important
in rethinking
work of the FTC
resolutions
promote
more holistic use of
authorities
With these
the FTC can better utilize
limited resources and
fix
market structures | 268 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Recent streamlining is improving resource allocation\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "The F\nT\nC\nvoted to approve a series of resolutions authorizing investigations into key law enforcement priorities for the next decade. Specifically, the resolutions direct agency staff to use “compulsory process,” such as subpoenas, to investigate\nenforcement priorities\nthe resolutions\nrepresent an important step in rethinking the work of the FTC\ncross-agency, investigatory resolutions will promote a more holistic use of the FTC’s enforcement authorities to stop bad actors\nThe actions taken\nwill\nallow staff to use compulsory process to investigate both proposed mergers and consummated mergers\nWith these in place, the FTC can better utilize its limited resources and move forward in earnest to fix\nmarket structures",
"role": "user"
},
{
"content": "The F\nT\nC\napprove\nresolutions authorizing\nkey\nenforcement priorities\nimportant\nin rethinking\nwork of the FTC\nresolutions\npromote\nmore holistic use of\nauthorities\nWith these\nthe FTC can better utilize\nlimited resources and\nfix\nmarket structures",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking. | In response to modern business practices, commentators are calling on the Federal Trade Commission's ("FTC" or "Commission") Bureau of Competition to adopt two previously rejected forms of regulation.' Most prominently, some argue that the FTC should ask courts to expand section 5 of the Federal Trade Commission Act ("FTC Act")2 to enjoin anti-competitive practices not currently prohibited by the Sherman Act.3 More tentatively, others argue that the FTC should use notice-and-comment rulemaking to regulate competition.4 Notably, both of these proposals lack an in-depth consideration of the role of Chevron deference.5 Advocates of expanding section 5 liability neglect to explore how the FTC can use Chevron deference to do so; advocates of rulemaking rarely look to the consequences of these formal interpretations on future litigation. With these possibilities in mind, this Note proposes that the FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation.
More than a pragmatic litigation strategy, this approach will allow the FTC to finally fulfill the competing ideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy run by, as President Woodrow Wilson worried, a "smug lot of experts. ''6 Regrettably, in the rollback of antitrust enforcement since the 1980s, the FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency, which acts parallel to the Department of Justice ("DOJ") Antitrust Division.7 As a result of this enforcement-only approach, generalist courts have been deprived of the FTC's norm-creating powers and have at times used incorrect presumptions to stymie FTC litigation.8 Importantly, these losses do not mean that the FTC should turn to regulation via technocratic rules: economic and political considerations urge case-by-case adjudication. They do suggest, however, that a more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in future litigation. That is, by using the Chevron framework, the FTC can create an antitrust regime capable of regulating ever-changing business practices. Simply stated, Chevron deference offers the FTC a route between technocracy and simple enforcement. | In response to
business practices, commentators are calling on the
FTC
Bureau of Competition to adopt
previously rejected forms of regulation.'
some argue
that the FTC should use notice-and-comment rulemaking to regulate competition
these proposals lack an in-depth consideration of
Chevron deference
the FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation
More than a pragmatic litigation strategy, this
will allow the FTC to finally fulfill the
ideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy
Regrettably, in the rollback of antitrust enforcement
the FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency
As a result of this
approach, generalist courts have been deprived of the FTC's norm-creating powers and have
used incorrect presumptions to stymie FTC litigation
economic and political considerations urge case-by-case adjudication
however
a more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in
litigation
by using the Chevron framework
the FTC can create an antitrust regime capable of regulating ever-changing business practices
Chevron deference offers the FTC a route between technocracy and simple enforcement | 269 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking.",
"role": "system"
},
{
"content": "In response to modern business practices, commentators are calling on the Federal Trade Commission's (\"FTC\" or \"Commission\") Bureau of Competition to adopt two previously rejected forms of regulation.' Most prominently, some argue that the FTC should ask courts to expand section 5 of the Federal Trade Commission Act (\"FTC Act\")2 to enjoin anti-competitive practices not currently prohibited by the Sherman Act.3 More tentatively, others argue that the FTC should use notice-and-comment rulemaking to regulate competition.4 Notably, both of these proposals lack an in-depth consideration of the role of Chevron deference.5 Advocates of expanding section 5 liability neglect to explore how the FTC can use Chevron deference to do so; advocates of rulemaking rarely look to the consequences of these formal interpretations on future litigation. With these possibilities in mind, this Note proposes that the FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation.\nMore than a pragmatic litigation strategy, this approach will allow the FTC to finally fulfill the competing ideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy run by, as President Woodrow Wilson worried, a \"smug lot of experts. ''6 Regrettably, in the rollback of antitrust enforcement since the 1980s, the FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency, which acts parallel to the Department of Justice (\"DOJ\") Antitrust Division.7 As a result of this enforcement-only approach, generalist courts have been deprived of the FTC's norm-creating powers and have at times used incorrect presumptions to stymie FTC litigation.8 Importantly, these losses do not mean that the FTC should turn to regulation via technocratic rules: economic and political considerations urge case-by-case adjudication. They do suggest, however, that a more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in future litigation. That is, by using the Chevron framework, the FTC can create an antitrust regime capable of regulating ever-changing business practices. Simply stated, Chevron deference offers the FTC a route between technocracy and simple enforcement.",
"role": "user"
},
{
"content": "In response to\nbusiness practices, commentators are calling on the\nFTC\nBureau of Competition to adopt\npreviously rejected forms of regulation.'\nsome argue\nthat the FTC should use notice-and-comment rulemaking to regulate competition\nthese proposals lack an in-depth consideration of\nChevron deference\nthe FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation\nMore than a pragmatic litigation strategy, this\nwill allow the FTC to finally fulfill the\nideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy\nRegrettably, in the rollback of antitrust enforcement\nthe FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency\nAs a result of this\napproach, generalist courts have been deprived of the FTC's norm-creating powers and have\nused incorrect presumptions to stymie FTC litigation\neconomic and political considerations urge case-by-case adjudication\nhowever\na more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in\nlitigation\nby using the Chevron framework\nthe FTC can create an antitrust regime capable of regulating ever-changing business practices\nChevron deference offers the FTC a route between technocracy and simple enforcement",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking. | In response to
business practices, commentators are calling on the
FTC
Bureau of Competition to adopt
previously rejected forms of regulation.'
some argue
that the FTC should use notice-and-comment rulemaking to regulate competition
these proposals lack an in-depth consideration of
Chevron deference
the FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation
More than a pragmatic litigation strategy, this
will allow the FTC to finally fulfill the
ideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy
Regrettably, in the rollback of antitrust enforcement
the FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency
As a result of this
approach, generalist courts have been deprived of the FTC's norm-creating powers and have
used incorrect presumptions to stymie FTC litigation
economic and political considerations urge case-by-case adjudication
however
a more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in
litigation
by using the Chevron framework
the FTC can create an antitrust regime capable of regulating ever-changing business practices
Chevron deference offers the FTC a route between technocracy and simple enforcement | business practices
FTC
rejected
regulation
use notice-and-comment rulemaking
regulate competition
Chevron deference
should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation
More
litigation
fulfill
design
norm-creators
limited
technocracy
neglected
carefully-crafted balance
turned
another enforcement agency
courts
deprived of
FTC's norm-creating powers
incorrect presumptions
stymie FTC litigation
case-by-case
however
effective
notice-and-comment rulemaking
attain Chevron deference in
litigation
Chevron framework
antitrust
regulating
changing
between technocracy and simple enforcement | 270 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking.",
"role": "system"
},
{
"content": "In response to\nbusiness practices, commentators are calling on the\nFTC\nBureau of Competition to adopt\npreviously rejected forms of regulation.'\nsome argue\nthat the FTC should use notice-and-comment rulemaking to regulate competition\nthese proposals lack an in-depth consideration of\nChevron deference\nthe FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation\nMore than a pragmatic litigation strategy, this\nwill allow the FTC to finally fulfill the\nideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy\nRegrettably, in the rollback of antitrust enforcement\nthe FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency\nAs a result of this\napproach, generalist courts have been deprived of the FTC's norm-creating powers and have\nused incorrect presumptions to stymie FTC litigation\neconomic and political considerations urge case-by-case adjudication\nhowever\na more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in\nlitigation\nby using the Chevron framework\nthe FTC can create an antitrust regime capable of regulating ever-changing business practices\nChevron deference offers the FTC a route between technocracy and simple enforcement",
"role": "user"
},
{
"content": "business practices\nFTC\nrejected\nregulation\nuse notice-and-comment rulemaking\nregulate competition\nChevron deference\nshould promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation\nMore\nlitigation\nfulfill\ndesign\nnorm-creators\nlimited\ntechnocracy\nneglected\ncarefully-crafted balance\nturned\nanother enforcement agency\ncourts\ndeprived of\nFTC's norm-creating powers\nincorrect presumptions\nstymie FTC litigation\ncase-by-case\nhowever\neffective\nnotice-and-comment rulemaking\nattain Chevron deference in\nlitigation\nChevron framework\nantitrust\nregulating\nchanging\nbetween technocracy and simple enforcement",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking.
Highlight Ratio: 0.3 | In response to
business practices, commentators are calling on the
FTC
Bureau of Competition to adopt
previously rejected forms of regulation.'
some argue
that the FTC should use notice-and-comment rulemaking to regulate competition
these proposals lack an in-depth consideration of
Chevron deference
the FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation
More than a pragmatic litigation strategy, this
will allow the FTC to finally fulfill the
ideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy
Regrettably, in the rollback of antitrust enforcement
the FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency
As a result of this
approach, generalist courts have been deprived of the FTC's norm-creating powers and have
used incorrect presumptions to stymie FTC litigation
economic and political considerations urge case-by-case adjudication
however
a more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in
litigation
by using the Chevron framework
the FTC can create an antitrust regime capable of regulating ever-changing business practices
Chevron deference offers the FTC a route between technocracy and simple enforcement | FTC should use notice-and-comment rulemaking to regulate competition
for the
purpose of attaining Chevron
in antitrust
this
allow
FTC to
fulfill
its
design
to be
norm-creators
FTC and courts have neglected this
balance and
turned
FTC into another enforcement agency
As a result
courts have been deprived of
FTC's norm-creating powers and
stymie FTC litigation
more effective enforcement
can be achieved
by using
Chevron
FTC can create
antitrust
capable of regulating
changing
practices | 271 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Solves the case AND its claim of authority forces courts to accord Chevron deference to FTC rulemaking.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "In response to\nbusiness practices, commentators are calling on the\nFTC\nBureau of Competition to adopt\npreviously rejected forms of regulation.'\nsome argue\nthat the FTC should use notice-and-comment rulemaking to regulate competition\nthese proposals lack an in-depth consideration of\nChevron deference\nthe FTC should promulgate notice-and-comment regulations for the express purpose of attaining Chevron deference in antitrust litigation\nMore than a pragmatic litigation strategy, this\nwill allow the FTC to finally fulfill the\nideals at the heart of its antitrust mandate. The FTC was designed to be an agency of norm-creators. Its powers and structure were also explicitly limited, however, to prevent it from becoming a technocracy\nRegrettably, in the rollback of antitrust enforcement\nthe FTC and courts have neglected this carefully-crafted balance and have turned the FTC into another enforcement agency\nAs a result of this\napproach, generalist courts have been deprived of the FTC's norm-creating powers and have\nused incorrect presumptions to stymie FTC litigation\neconomic and political considerations urge case-by-case adjudication\nhowever\na more effective enforcement regime can be achieved if the FTC uses notice-and-comment rulemaking to attain Chevron deference in\nlitigation\nby using the Chevron framework\nthe FTC can create an antitrust regime capable of regulating ever-changing business practices\nChevron deference offers the FTC a route between technocracy and simple enforcement",
"role": "user"
},
{
"content": "FTC should use notice-and-comment rulemaking to regulate competition\nfor the\npurpose of attaining Chevron\nin antitrust\nthis\nallow\nFTC to\nfulfill\nits\ndesign\nto be\nnorm-creators\nFTC and courts have neglected this\nbalance and\nturned\nFTC into another enforcement agency\nAs a result\ncourts have been deprived of\nFTC's norm-creating powers and\nstymie FTC litigation\nmore effective enforcement\ncan be achieved\nby using\nChevron\nFTC can create\nantitrust\ncapable of regulating\nchanging\npractices",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger. | VI. CONCLUSION
The FTC should assert, and courts should grant, Chevron deference to protect FTC expert decisions from judicial meddling and to ultimately establish the technocratic institutional structure that Congress sought in 1914. Congress may have felt that antitrust law had reached a "technical" tipping point in 1914 when it founded the FTC to be "more competent to deal with the practical ... dissolution of [business] combinations than any court or Attorney General could be." 219 Yet contrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role that it had prior to the FTC Act's passing. It is true that courts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron deference if it used its rulemaking power and bolstered its expertise in key areas.
In closing, it seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye. Perhaps antitrust is merely at an interim point; despite not adhering strictly to Congress's 1914 desires, courts have gradually been moving in a technocratic direction with the modern economy. For the sake of this modern economy, one hopes the FTC will be proactive, seize on this technocratic shift, invoke Chevron deference, and courts will trust the FTC's relative superiority. | The FTC should assert
Chevron deference to protect FTC expert decisions from judicial meddling and
establish the technocratic institutional structure that Congress sought in 1914
antitrust law
reached a "technical" tipping point in 1914
contrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role
it had prior to the FTC Act's passing
courts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron
if it used its rulemaking power and bolstered its expertise
it seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye
For the sake of this modern economy, one hopes the FTC will be proactive
seize on this technocratic shift, invoke Chevron
and courts will trust the FTC's relative superiority | 272 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger.",
"role": "system"
},
{
"content": "VI. CONCLUSION\nThe FTC should assert, and courts should grant, Chevron deference to protect FTC expert decisions from judicial meddling and to ultimately establish the technocratic institutional structure that Congress sought in 1914. Congress may have felt that antitrust law had reached a \"technical\" tipping point in 1914 when it founded the FTC to be \"more competent to deal with the practical ... dissolution of [business] combinations than any court or Attorney General could be.\" 219 Yet contrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role that it had prior to the FTC Act's passing. It is true that courts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron deference if it used its rulemaking power and bolstered its expertise in key areas.\nIn closing, it seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye. Perhaps antitrust is merely at an interim point; despite not adhering strictly to Congress's 1914 desires, courts have gradually been moving in a technocratic direction with the modern economy. For the sake of this modern economy, one hopes the FTC will be proactive, seize on this technocratic shift, invoke Chevron deference, and courts will trust the FTC's relative superiority.",
"role": "user"
},
{
"content": "The FTC should assert\nChevron deference to protect FTC expert decisions from judicial meddling and\nestablish the technocratic institutional structure that Congress sought in 1914\nantitrust law\nreached a \"technical\" tipping point in 1914\ncontrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role\nit had prior to the FTC Act's passing\ncourts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron\nif it used its rulemaking power and bolstered its expertise\nit seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye\nFor the sake of this modern economy, one hopes the FTC will be proactive\nseize on this technocratic shift, invoke Chevron\nand courts will trust the FTC's relative superiority",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger. | The FTC should assert
Chevron deference to protect FTC expert decisions from judicial meddling and
establish the technocratic institutional structure that Congress sought in 1914
antitrust law
reached a "technical" tipping point in 1914
contrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role
it had prior to the FTC Act's passing
courts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron
if it used its rulemaking power and bolstered its expertise
it seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye
For the sake of this modern economy, one hopes the FTC will be proactive
seize on this technocratic shift, invoke Chevron
and courts will trust the FTC's relative superiority | should assert
Chevron deference to protect FTC expert decisions from judicial meddling
technocratic institutional structure that Congress sought in 1914
"technical" tipping point
judiciary
same institutional role
have
grown
procedural barriers
substantive standards
yet to grant the FTC Chevron deference
FTC's
experts
more likely
used its rulemaking power
bolstered its expertise
doubtful
less technical, less complex, or less opaque
more
more
FTC
proactive
seize
invoke Chevron
courts
trust the FTC's relative superiority | 273 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger.",
"role": "system"
},
{
"content": "The FTC should assert\nChevron deference to protect FTC expert decisions from judicial meddling and\nestablish the technocratic institutional structure that Congress sought in 1914\nantitrust law\nreached a \"technical\" tipping point in 1914\ncontrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role\nit had prior to the FTC Act's passing\ncourts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron\nif it used its rulemaking power and bolstered its expertise\nit seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye\nFor the sake of this modern economy, one hopes the FTC will be proactive\nseize on this technocratic shift, invoke Chevron\nand courts will trust the FTC's relative superiority",
"role": "user"
},
{
"content": "should assert\nChevron deference to protect FTC expert decisions from judicial meddling\ntechnocratic institutional structure that Congress sought in 1914\n\"technical\" tipping point\njudiciary\nsame institutional role\nhave\ngrown\nprocedural barriers\nsubstantive standards\nyet to grant the FTC Chevron deference\nFTC's\nexperts\nmore likely\nused its rulemaking power\nbolstered its expertise\ndoubtful\nless technical, less complex, or less opaque\nmore\nmore\nFTC\nproactive\nseize\ninvoke Chevron\ncourts\ntrust the FTC's relative superiority",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger.
Highlight Ratio: 0.3 | The FTC should assert
Chevron deference to protect FTC expert decisions from judicial meddling and
establish the technocratic institutional structure that Congress sought in 1914
antitrust law
reached a "technical" tipping point in 1914
contrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role
it had prior to the FTC Act's passing
courts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron
if it used its rulemaking power and bolstered its expertise
it seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye
For the sake of this modern economy, one hopes the FTC will be proactive
seize on this technocratic shift, invoke Chevron
and courts will trust the FTC's relative superiority | FTC should assert
deference to protect FTC
decisions from
meddling
the judiciary has yet to grant
Chevron
to protect
FTC's role as experts
would
more likely
receive Chevron
if it used
rulemaking
it seems doubtful
antitrust will get less technical
complex, or
opaque. The economy will
grow more specialized
economics will require a more discerning eye
one hopes
FTC will
seize
this
shift, invoke Chevron
and courts will trust
FTC's
superiority | 274 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: 2) EXPERTISE---Courts are incompetent due to the complexity of antitrust enforcement. Granting the FTC Chevron deference makes enforcement stronger.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "The FTC should assert\nChevron deference to protect FTC expert decisions from judicial meddling and\nestablish the technocratic institutional structure that Congress sought in 1914\nantitrust law\nreached a \"technical\" tipping point in 1914\ncontrary to Congressional desires, the judiciary has charted a path to maintain the same institutional role\nit had prior to the FTC Act's passing\ncourts have recognized their institutional foibles as antitrust law has grown more complex; indeed, courts have erected procedural barriers and created heightened substantive standards to protect themselves from false positive rulings. But the judiciary has yet to grant the FTC Chevron deference to protect the FTC's role as experts and policymakers. The FTC would be more likely to receive Chevron\nif it used its rulemaking power and bolstered its expertise\nit seems doubtful that antitrust will get less technical, less complex, or less opaque. The economy will likely grow more specialized, and antitrust economics will require a more discerning eye\nFor the sake of this modern economy, one hopes the FTC will be proactive\nseize on this technocratic shift, invoke Chevron\nand courts will trust the FTC's relative superiority",
"role": "user"
},
{
"content": "FTC should assert\ndeference to protect FTC\ndecisions from\nmeddling\nthe judiciary has yet to grant\nChevron\nto protect\nFTC's role as experts\nwould\nmore likely\nreceive Chevron\nif it used\nrulemaking\nit seems doubtful\nantitrust will get less technical\ncomplex, or\nopaque. The economy will\ngrow more specialized\neconomics will require a more discerning eye\none hopes\nFTC will\nseize\nthis\nshift, invoke Chevron\nand courts will trust\nFTC's\nsuperiority",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules. | And third, rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in on it, granting the rule greater legitimacy.49 APA procedures require that an agency provide the public with meaningful opportunity to comment on the rule’s content through the submission of written “data, views, or arguments.”50 The agency must then consider and address all submitted comments before issuing the final rule. If an agency adopts a rule without observing these procedures, a court may strike down the rule.51
This process is far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent-setting rules and standards, the Commission can put forth rules after considering a comprehensive set of information and analysis.52 Notably, this would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and industry practices over the long term and outside the context of litigation.53 Drawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices. | rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in
granting the rule greater legitimacy
APA procedures require
an agency provide
meaningful opportunity to comment on the rule’s content through
submission of written “data, views, or arguments.”
The agency must then consider and address all submitted comments
This
is far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent
the Commission can put forth rules after considering a comprehensive set of information
this would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and
practices over the long term and outside the context of litigation
Drawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices. | 275 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules.",
"role": "system"
},
{
"content": "And third, rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in on it, granting the rule greater legitimacy.49 APA procedures require that an agency provide the public with meaningful opportunity to comment on the rule’s content through the submission of written “data, views, or arguments.”50 The agency must then consider and address all submitted comments before issuing the final rule. If an agency adopts a rule without observing these procedures, a court may strike down the rule.51\nThis process is far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent-setting rules and standards, the Commission can put forth rules after considering a comprehensive set of information and analysis.52 Notably, this would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and industry practices over the long term and outside the context of litigation.53 Drawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices.",
"role": "user"
},
{
"content": "rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in\ngranting the rule greater legitimacy\nAPA procedures require\nan agency provide\nmeaningful opportunity to comment on the rule’s content through\nsubmission of written “data, views, or arguments.”\nThe agency must then consider and address all submitted comments\nThis\nis far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent\nthe Commission can put forth rules after considering a comprehensive set of information\nthis would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and\npractices over the long term and outside the context of litigation\nDrawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules. | rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in
granting the rule greater legitimacy
APA procedures require
an agency provide
meaningful opportunity to comment on the rule’s content through
submission of written “data, views, or arguments.”
The agency must then consider and address all submitted comments
This
is far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent
the Commission can put forth rules after considering a comprehensive set of information
this would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and
practices over the long term and outside the context of litigation
Drawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices. | transparent
participatory
weigh in
legitimacy
comment
content
all submitted comments
far more participatory
judges
trial record
Commission
consider
comprehensive
info
own info
advantage
aggregate information
market trends
practices
long term
outside
context of litigation
better reflect empirical realities
keep pace with evolving business practices. | 276 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules.",
"role": "system"
},
{
"content": "rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in\ngranting the rule greater legitimacy\nAPA procedures require\nan agency provide\nmeaningful opportunity to comment on the rule’s content through\nsubmission of written “data, views, or arguments.”\nThe agency must then consider and address all submitted comments\nThis\nis far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent\nthe Commission can put forth rules after considering a comprehensive set of information\nthis would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and\npractices over the long term and outside the context of litigation\nDrawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices.",
"role": "user"
},
{
"content": "transparent\nparticipatory\nweigh in\nlegitimacy\ncomment\ncontent\nall submitted comments\nfar more participatory\njudges\ntrial record\nCommission\nconsider\ncomprehensive\ninfo\nown info\nadvantage\naggregate information\nmarket trends\npractices\nlong term\noutside\ncontext of litigation\nbetter reflect empirical realities\nkeep pace with evolving business practices.",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules.
Highlight Ratio: 0.4 | rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in
granting the rule greater legitimacy
APA procedures require
an agency provide
meaningful opportunity to comment on the rule’s content through
submission of written “data, views, or arguments.”
The agency must then consider and address all submitted comments
This
is far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent
the Commission can put forth rules after considering a comprehensive set of information
this would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and
practices over the long term and outside the context of litigation
Drawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices. | rulemaking would
establish
a transparent
participatory process
APA procedures require
an
opportunity to comment
The agency must
address
submitted comments
This
is
more participatory
judges
are confined to the trial record
the Commission can
consider
comprehensive
info
this
allow
FTC to draw on its
info
advantage
ability to
aggregate
and
study
trends and
practices over the long term and outside
litigation | 277 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: 3) PARTICIPATION---the rulemaking process is transparent and open to stakeholder input, which results in substantively superior rules.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "rulemaking would enable the Commission to establish rules through a transparent and participatory process, ensuring that everyone who may be affected by a new rule has the opportunity to weigh in\ngranting the rule greater legitimacy\nAPA procedures require\nan agency provide\nmeaningful opportunity to comment on the rule’s content through\nsubmission of written “data, views, or arguments.”\nThe agency must then consider and address all submitted comments\nThis\nis far more participatory than adjudication. Unlike judges, who are confined to the trial record when developing precedent\nthe Commission can put forth rules after considering a comprehensive set of information\nthis would also allow the FTC to draw on its own informational advantage—namely, its ability to collect and aggregate information and to study market trends and\npractices over the long term and outside the context of litigation\nDrawing on this expertise to develop rules will help antitrust enforcement and policymaking better reflect empirical realities and better keep pace with evolving business practices.",
"role": "user"
},
{
"content": "rulemaking would\nestablish\na transparent\nparticipatory process\nAPA procedures require\nan\nopportunity to comment\nThe agency must\naddress\nsubmitted comments\nThis\nis\nmore participatory\njudges\nare confined to the trial record\nthe Commission can\nconsider\ncomprehensive\ninfo\nthis\nallow\nFTC to draw on its\ninfo\nadvantage\nability to\naggregate\nand\nstudy\ntrends and\npractices over the long term and outside\nlitigation",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban. | “Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting. As former Commissioner Elman rightly noted, rulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions.”41 Rules come from courts, legislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them.42 The merger guidelines stipulate the analytical framework that the agencies rely on to enforce the merger law. Agency rulemaking could do the same for “unfair methods of competition.” | “Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting
rulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions
Rules come from courts
legislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them
stipulate the analytical framework
agencies rely on to enforce the
law. Agency rulemaking could do the same for “unfair methods of competition.” | 278 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban.",
"role": "system"
},
{
"content": "“Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting. As former Commissioner Elman rightly noted, rulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions.”41 Rules come from courts, legislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them.42 The merger guidelines stipulate the analytical framework that the agencies rely on to enforce the merger law. Agency rulemaking could do the same for “unfair methods of competition.”",
"role": "user"
},
{
"content": "“Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting\nrulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions\nRules come from courts\nlegislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them\nstipulate the analytical framework\nagencies rely on to enforce the\nlaw. Agency rulemaking could do the same for “unfair methods of competition.”",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban. | “Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting
rulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions
Rules come from courts
legislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them
stipulate the analytical framework
agencies rely on to enforce the
law. Agency rulemaking could do the same for “unfair methods of competition.” | idea
gov
inflexible prescription upon the marketplace
not what we are suggesting
also
standards, guidelines, pointers, criteria, or presumptions
courts
legislative
agencies
promulgated as
merger guidelines
as
courts
analytical framework
could do the same
unfair methods of competition | 279 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban.",
"role": "system"
},
{
"content": "“Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting\nrulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions\nRules come from courts\nlegislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them\nstipulate the analytical framework\nagencies rely on to enforce the\nlaw. Agency rulemaking could do the same for “unfair methods of competition.”",
"role": "user"
},
{
"content": "idea\ngov\ninflexible prescription upon the marketplace\nnot what we are suggesting\nalso\nstandards, guidelines, pointers, criteria, or presumptions\ncourts\nlegislative\nagencies\npromulgated as\nmerger guidelines\nas\ncourts\nanalytical framework\ncould do the same\nunfair methods of competition",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban.
Highlight Ratio: 0.6 | “Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting
rulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions
Rules come from courts
legislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them
stipulate the analytical framework
agencies rely on to enforce the
law. Agency rulemaking could do the same for “unfair methods of competition.” | “Rulemaking”
evokes the idea of
some inflexible prescription
This is not what we are suggesting
rulemaking can also be
standards, guidelines, pointers, criteria, or presumptions
While
not promulgated as
rules
the merger guidelines
came to serve as rules
stipulate the analytical framework
to enforce
law
rulemaking could do the same for “unfair
competition | 280 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Rulemaking can create standards for the application of rule of reason determinations. Doesn’t have to be a blanket ban.\n\nHighlight Ratio: 0.6",
"role": "system"
},
{
"content": "“Rulemaking” often evokes the idea of government imposing some inflexible prescription upon the marketplace. This is not what we are suggesting\nrulemaking can also be related to “standards, guidelines, pointers, criteria, or presumptions\nRules come from courts\nlegislative bodies, and agencies. While they were not promulgated as agency rules, certain elements of the merger guidelines eventually came to serve as rules once courts adopted them\nstipulate the analytical framework\nagencies rely on to enforce the\nlaw. Agency rulemaking could do the same for “unfair methods of competition.”",
"role": "user"
},
{
"content": "“Rulemaking”\nevokes the idea of\nsome inflexible prescription\nThis is not what we are suggesting\nrulemaking can also be\nstandards, guidelines, pointers, criteria, or presumptions\nWhile\nnot promulgated as\nrules\nthe merger guidelines\ncame to serve as rules\nstipulate the analytical framework\nto enforce\nlaw\nrulemaking could do the same for “unfair\ncompetition",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time | B. The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron
The above limitations take on particular salience within antitrust jurisprudence. Courts often decide antitrust cases not based on certainty, but by referring to a legal presumption about market structure. The advantage of these presumptions is that they obviate the need for an extended analysis.1 28 Some of these presumptions are simply about the likelihood of particular conduct actually occurring. For example, "predatory pricing schemes are rarely tried, and even more rarely successful. 129 Others are fundamental presumptions regarding the merit of the conduct: e.g., the standard for per se liability.1 30 While it makes sense to use these presumptions, the basis on which courts actually form hese broad determinations is unclear. Are courts limited to information provided by litigants? Or do they leverage a wider scope of judicial understanding?
With these institutional differences in mind, this Note advocates that the FTC take advantage of Chevron deference to form, with judicial oversight and consent, the presumptions needed for antitrust adjudication. Crucially, this form of regulation is of a different style than traditional notice-and-comment regulation. Often, regulations set out standards to which regulated parties must adhere. In comparison, these proposed regulations are legal presumptions for the judiciary to use while judging claims of "unfair business practices." Thus, these rules differ from direct regulation of business, because they would not directly prohibit certain conduct, but would instead set out the legal framework for a reviewing court to use. In this sense, the FTC has the ability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis.
An instructive use of this style of regulation occurred in 1991 with the National Labor Relations Board's ("NLRB") promulgation of 29 C.F.R. § 103.30. There, the NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals.1 3 1 Specifically, the regulation created the legal presumption that, absent "extraordinary circumstances," there were only eight possible collective bargaining units in acute care hospitals. 132 In limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent and create a legal presumption for courts to employ. After promulgation, this rule was challenged and a unanimous Supreme Court upheld the regulation partly based on Chevron deference. 33 Notably, this regulation did not turn litigation into a simple application of predetermined values (as the rules in Vermont Yankee did). 3'4 It simply set the presumption for generalist courts to deploy in deciding the existence of bargaining units.
Turning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.'135 Alternatively, the FTC could expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine.13 6 In assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority. As previously discussed, the interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust norms.1 37 The inner limitation prevents Chevron deference from extending to common law legal terms. The FTC cannot create the norms of the Sherman Act, and therefore cannot receive Chevron deference for its interpretations of those parts of section 5 that simply incorporate these principles. As a practical matter, this often amounts to a procedural drafting requirement rather than a substantive limitation.1 38 The FTC cannot simply conclude that a "monopoly is X." The FTC can, however, conclude that an "unfair section 5 monopoly is X." | The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron
this Note advocates that the FTC take advantage of Chevron deference to form
presumptions needed for antitrust adjudication
this form of regulation is of a different style than traditional notice-and-comment
Often, regulations set out standards
In comparison, these
regulations are legal presumptions
to use while judging claims of "unfair business practices." Thus, these rules differ from direct regulation
because they would not directly prohibit
conduct, but would instead set out the legal framework for a reviewing court to use
this
has
ability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis
An instructive use of this style of regulation occurred
with the
NLRB
the NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals
Specifically, the regulation created the legal presumption that, absent "extraordinary circumstances," there were only eight possible collective bargaining units in acute care hospitals
In limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent
After promulgation, this rule was challenged and a unanimous
Court upheld the regulation
based on Chevron
Turning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.
Alternatively
could expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine
In assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority
the interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust
The inner limitation prevents Chevron deference from extending to common law legal terms
The FTC cannot simply conclude that a "monopoly is X." The FTC can, however, conclude that an "unfair section 5 monopoly is X." | 281 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time",
"role": "system"
},
{
"content": "B. The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron\nThe above limitations take on particular salience within antitrust jurisprudence. Courts often decide antitrust cases not based on certainty, but by referring to a legal presumption about market structure. The advantage of these presumptions is that they obviate the need for an extended analysis.1 28 Some of these presumptions are simply about the likelihood of particular conduct actually occurring. For example, \"predatory pricing schemes are rarely tried, and even more rarely successful. 129 Others are fundamental presumptions regarding the merit of the conduct: e.g., the standard for per se liability.1 30 While it makes sense to use these presumptions, the basis on which courts actually form hese broad determinations is unclear. Are courts limited to information provided by litigants? Or do they leverage a wider scope of judicial understanding?\nWith these institutional differences in mind, this Note advocates that the FTC take advantage of Chevron deference to form, with judicial oversight and consent, the presumptions needed for antitrust adjudication. Crucially, this form of regulation is of a different style than traditional notice-and-comment regulation. Often, regulations set out standards to which regulated parties must adhere. In comparison, these proposed regulations are legal presumptions for the judiciary to use while judging claims of \"unfair business practices.\" Thus, these rules differ from direct regulation of business, because they would not directly prohibit certain conduct, but would instead set out the legal framework for a reviewing court to use. In this sense, the FTC has the ability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis.\nAn instructive use of this style of regulation occurred in 1991 with the National Labor Relations Board's (\"NLRB\") promulgation of 29 C.F.R. § 103.30. There, the NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals.1 3 1 Specifically, the regulation created the legal presumption that, absent \"extraordinary circumstances,\" there were only eight possible collective bargaining units in acute care hospitals. 132 In limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent and create a legal presumption for courts to employ. After promulgation, this rule was challenged and a unanimous Supreme Court upheld the regulation partly based on Chevron deference. 33 Notably, this regulation did not turn litigation into a simple application of predetermined values (as the rules in Vermont Yankee did). 3'4 It simply set the presumption for generalist courts to deploy in deciding the existence of bargaining units.\nTurning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.'135 Alternatively, the FTC could expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine.13 6 In assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority. As previously discussed, the interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust norms.1 37 The inner limitation prevents Chevron deference from extending to common law legal terms. The FTC cannot create the norms of the Sherman Act, and therefore cannot receive Chevron deference for its interpretations of those parts of section 5 that simply incorporate these principles. As a practical matter, this often amounts to a procedural drafting requirement rather than a substantive limitation.1 38 The FTC cannot simply conclude that a \"monopoly is X.\" The FTC can, however, conclude that an \"unfair section 5 monopoly is X.\"",
"role": "user"
},
{
"content": "The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron\nthis Note advocates that the FTC take advantage of Chevron deference to form\npresumptions needed for antitrust adjudication\nthis form of regulation is of a different style than traditional notice-and-comment\nOften, regulations set out standards\nIn comparison, these\nregulations are legal presumptions\nto use while judging claims of \"unfair business practices.\" Thus, these rules differ from direct regulation\nbecause they would not directly prohibit\nconduct, but would instead set out the legal framework for a reviewing court to use\nthis\nhas\nability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis\nAn instructive use of this style of regulation occurred\nwith the\nNLRB\nthe NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals\nSpecifically, the regulation created the legal presumption that, absent \"extraordinary circumstances,\" there were only eight possible collective bargaining units in acute care hospitals\nIn limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent\nAfter promulgation, this rule was challenged and a unanimous\nCourt upheld the regulation\nbased on Chevron\nTurning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.\nAlternatively\ncould expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine\nIn assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority\nthe interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust\nThe inner limitation prevents Chevron deference from extending to common law legal terms\nThe FTC cannot simply conclude that a \"monopoly is X.\" The FTC can, however, conclude that an \"unfair section 5 monopoly is X.\"",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time | The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron
this Note advocates that the FTC take advantage of Chevron deference to form
presumptions needed for antitrust adjudication
this form of regulation is of a different style than traditional notice-and-comment
Often, regulations set out standards
In comparison, these
regulations are legal presumptions
to use while judging claims of "unfair business practices." Thus, these rules differ from direct regulation
because they would not directly prohibit
conduct, but would instead set out the legal framework for a reviewing court to use
this
has
ability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis
An instructive use of this style of regulation occurred
with the
NLRB
the NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals
Specifically, the regulation created the legal presumption that, absent "extraordinary circumstances," there were only eight possible collective bargaining units in acute care hospitals
In limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent
After promulgation, this rule was challenged and a unanimous
Court upheld the regulation
based on Chevron
Turning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.
Alternatively
could expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine
In assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority
the interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust
The inner limitation prevents Chevron deference from extending to common law legal terms
The FTC cannot simply conclude that a "monopoly is X." The FTC can, however, conclude that an "unfair section 5 monopoly is X." | Expertise-Driven
Further Justifies the FTC's Use of Chevron
take advantage of Chevron deference
form
presumptions
antitrust adjudication
different style
traditional notice-and-comment
reg
s
standards
comparison
these
legal presumptions
while judging claims
differ
direct
not directly prohibit
instead
legal framework
review
continually-adapting section 5 jurisprudence free
ossifying effects
NLRB
litigation
legal presumption
eight possible collective bargaining units
overrule conflicting precedent
challenged
unanimous
Court upheld the regulation
based on Chevron
promulgate fine-grained presumptions that will be applied formulaically
Alternatively
expand liability
frameworks
assessing
essential facilities doctrine
ability
form
permitted
can
conclude that an "unfair section 5 monopoly is X." | 282 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time",
"role": "system"
},
{
"content": "The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron\nthis Note advocates that the FTC take advantage of Chevron deference to form\npresumptions needed for antitrust adjudication\nthis form of regulation is of a different style than traditional notice-and-comment\nOften, regulations set out standards\nIn comparison, these\nregulations are legal presumptions\nto use while judging claims of \"unfair business practices.\" Thus, these rules differ from direct regulation\nbecause they would not directly prohibit\nconduct, but would instead set out the legal framework for a reviewing court to use\nthis\nhas\nability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis\nAn instructive use of this style of regulation occurred\nwith the\nNLRB\nthe NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals\nSpecifically, the regulation created the legal presumption that, absent \"extraordinary circumstances,\" there were only eight possible collective bargaining units in acute care hospitals\nIn limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent\nAfter promulgation, this rule was challenged and a unanimous\nCourt upheld the regulation\nbased on Chevron\nTurning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.\nAlternatively\ncould expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine\nIn assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority\nthe interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust\nThe inner limitation prevents Chevron deference from extending to common law legal terms\nThe FTC cannot simply conclude that a \"monopoly is X.\" The FTC can, however, conclude that an \"unfair section 5 monopoly is X.\"",
"role": "user"
},
{
"content": "Expertise-Driven\nFurther Justifies the FTC's Use of Chevron\ntake advantage of Chevron deference\nform\npresumptions\nantitrust adjudication\ndifferent style\ntraditional notice-and-comment\nreg\ns\nstandards\ncomparison\nthese\nlegal presumptions\nwhile judging claims\ndiffer\ndirect\nnot directly prohibit\ninstead\nlegal framework\nreview\ncontinually-adapting section 5 jurisprudence free\nossifying effects\nNLRB\nlitigation\nlegal presumption\neight possible collective bargaining units\noverrule conflicting precedent\nchallenged\nunanimous\nCourt upheld the regulation\nbased on Chevron\npromulgate fine-grained presumptions that will be applied formulaically\nAlternatively\nexpand liability\nframeworks\nassessing\nessential facilities doctrine\nability\nform\npermitted\ncan\nconclude that an \"unfair section 5 monopoly is X.\"",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time
Highlight Ratio: 0.2 | The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron
this Note advocates that the FTC take advantage of Chevron deference to form
presumptions needed for antitrust adjudication
this form of regulation is of a different style than traditional notice-and-comment
Often, regulations set out standards
In comparison, these
regulations are legal presumptions
to use while judging claims of "unfair business practices." Thus, these rules differ from direct regulation
because they would not directly prohibit
conduct, but would instead set out the legal framework for a reviewing court to use
this
has
ability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis
An instructive use of this style of regulation occurred
with the
NLRB
the NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals
Specifically, the regulation created the legal presumption that, absent "extraordinary circumstances," there were only eight possible collective bargaining units in acute care hospitals
In limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent
After promulgation, this rule was challenged and a unanimous
Court upheld the regulation
based on Chevron
Turning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.
Alternatively
could expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine
In assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority
the interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust
The inner limitation prevents Chevron deference from extending to common law legal terms
The FTC cannot simply conclude that a "monopoly is X." The FTC can, however, conclude that an "unfair section 5 monopoly is X." | FTC
Chevron
form
presumptions
for
adjudication
this
is
a different style than traditional notice-and-comment
reg
s set
standards
to use while judging claims of "unfair
practices."
these
differ from direct regulation
they would not directly prohibit
but
instead set
the legal framework for
review
create a continually-adapting section 5 jurisprudence free of
ossifying effects
this
occurred
with
NLRB
was challenged and a unanimous
Court upheld
based on Chevron
section 5
can promulgate fine-grained presumptions
applied formulaically
Alternatively
expand liability by adopting
frameworks for assessing
e.g.,
essential facilities | 283 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Empirics from the NLRB prove that such ‘balancing test’ rulemaking can receive Chevron deference---AND, it can be flexible over time\n\nHighlight Ratio: 0.2",
"role": "system"
},
{
"content": "The Expertise-Driven Nature of Antitrust Jurisprudence Further Justifies the FTC's Use of Chevron\nthis Note advocates that the FTC take advantage of Chevron deference to form\npresumptions needed for antitrust adjudication\nthis form of regulation is of a different style than traditional notice-and-comment\nOften, regulations set out standards\nIn comparison, these\nregulations are legal presumptions\nto use while judging claims of \"unfair business practices.\" Thus, these rules differ from direct regulation\nbecause they would not directly prohibit\nconduct, but would instead set out the legal framework for a reviewing court to use\nthis\nhas\nability to create a continually-adapting section 5 jurisprudence free of the ossifying effects of stare decisis\nAn instructive use of this style of regulation occurred\nwith the\nNLRB\nthe NLRB promulgated its first rule seeking to cease the costly, frequent, and ineffective litigation aimed at determining collective bargaining units in hospitals\nSpecifically, the regulation created the legal presumption that, absent \"extraordinary circumstances,\" there were only eight possible collective bargaining units in acute care hospitals\nIn limiting the presumed form of bargaining units, the NLRB specifically intended to overrule conflicting precedent\nAfter promulgation, this rule was challenged and a unanimous\nCourt upheld the regulation\nbased on Chevron\nTurning to section 5, the FTC can promulgate fine-grained presumptions that will be applied formulaically; e.g., when evaluating predation claims, the appropriate marginal cost of commodity Y is X.\nAlternatively\ncould expand liability by adopting entire frameworks for assessing liability, e.g., the essential facilities doctrine\nIn assessing the merits of any of these uses, the fundamental considerations are that a presumption must (1) take advantage of the FTC's unique ability to form these presumptions, and (2) actually be within the FTC's delegated authority\nthe interaction of the Sherman Act with section 5 means that FTC Chevron space has an outer and an inner limitation. The outer limitation prevents section 5 from being interpreted to prevent conduct that is clearly permitted by antitrust\nThe inner limitation prevents Chevron deference from extending to common law legal terms\nThe FTC cannot simply conclude that a \"monopoly is X.\" The FTC can, however, conclude that an \"unfair section 5 monopoly is X.\"",
"role": "user"
},
{
"content": "FTC\nChevron\nform\npresumptions\nfor\nadjudication\nthis\nis\na different style than traditional notice-and-comment\nreg\ns set\nstandards\nto use while judging claims of \"unfair\npractices.\"\nthese\ndiffer from direct regulation\nthey would not directly prohibit\nbut\ninstead set\nthe legal framework for\nreview\ncreate a continually-adapting section 5 jurisprudence free of\nossifying effects\nthis\noccurred\nwith\nNLRB\nwas challenged and a unanimous\nCourt upheld\nbased on Chevron\nsection 5\ncan promulgate fine-grained presumptions\napplied formulaically\nAlternatively\nexpand liability by adopting\nframeworks for assessing\ne.g.,\nessential facilities",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule. | A. Section 5 is Precisely the Sort of Statute to Which Chevron Applies
As a threshold matter, Section 5 is precisely the sort of statute to which Chevron deference is meant to apply.167 At a mechanical level, Chevron instructs courts to first ask whether the meaning of the statute is clear.168 Both “unfair methods of competition” and “unfair or deceptive acts or practices” are inherently ambiguous; courts need not turn to historical documents to determine whether a specific meaning was intended by Congress or whether Congress clearly intended to delegate interpretive authority to the FTC. Nearly every word of the statute is rife with ambiguity: What is unfair? Unfair to whom? What is deceptive? What is a method? An act? A practice? What is competition? As the Court has noted, the standard is “by necessity, an elusive one.”169
Absent clarifying language in the statute itself, or in some cases references outside the statute that indicate contrary congressional intent,170 the ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron deference. The sole task of the courts is—or should be—to ensure that, whatever construction the FTC gives to Section 5, that construction is permissible within the boundaries of the statute.171
The argument for deference is even stronger when we consider outside references. The statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion to define the scope of Section 5 and, in particular, that the scope of Section 5 is broader than that of the antitrust laws.172 Section 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly;173 it was deliberately drafted with language that had not previously been considered by the courts.174 When the Court imposed an overly narrow construction on the statute in the 1950s, Congress amended the statute to overcome that narrowing interpretation.175
Section 5 is, thus, a case study in each of the four rationales for Chevron deference:176 congressional intent; agency expertise; concern about the courts’ limited political accountability as compared to Congress and its agencies; and the separation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference than Section 5. | As a threshold matter, Section 5 is precisely the sort of statute to which Chevron
is meant to apply
At a mechanical level, Chevron instructs
to first ask whether the
statute is clear
u
m
c
are inherently ambiguous
Nearly every word
is rife with ambiguity: What is unfair? Unfair to whom?
What is a method?
A practice? What is competition? As the Court has noted, the standard is “by necessity
elusive
Absent clarifying language in the statute itself
the ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron
The sole task of the courts is
to ensure that
construction is permissible
The statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion
Section 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly
When the Court imposed an overly narrow construction
Congress amended the statute to overcome that narrowing interpretation
Section 5 is, thus, a case study in each of the four rationales for Chevron deference
congressional intent; agency expertise; concern about
courts’ limited political accountability
and
separation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference | 284 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule.",
"role": "system"
},
{
"content": "A. Section 5 is Precisely the Sort of Statute to Which Chevron Applies \nAs a threshold matter, Section 5 is precisely the sort of statute to which Chevron deference is meant to apply.167 At a mechanical level, Chevron instructs courts to first ask whether the meaning of the statute is clear.168 Both “unfair methods of competition” and “unfair or deceptive acts or practices” are inherently ambiguous; courts need not turn to historical documents to determine whether a specific meaning was intended by Congress or whether Congress clearly intended to delegate interpretive authority to the FTC. Nearly every word of the statute is rife with ambiguity: What is unfair? Unfair to whom? What is deceptive? What is a method? An act? A practice? What is competition? As the Court has noted, the standard is “by necessity, an elusive one.”169 \nAbsent clarifying language in the statute itself, or in some cases references outside the statute that indicate contrary congressional intent,170 the ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron deference. The sole task of the courts is—or should be—to ensure that, whatever construction the FTC gives to Section 5, that construction is permissible within the boundaries of the statute.171 \nThe argument for deference is even stronger when we consider outside references. The statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion to define the scope of Section 5 and, in particular, that the scope of Section 5 is broader than that of the antitrust laws.172 Section 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly;173 it was deliberately drafted with language that had not previously been considered by the courts.174 When the Court imposed an overly narrow construction on the statute in the 1950s, Congress amended the statute to overcome that narrowing interpretation.175 \nSection 5 is, thus, a case study in each of the four rationales for Chevron deference:176 congressional intent; agency expertise; concern about the courts’ limited political accountability as compared to Congress and its agencies; and the separation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference than Section 5. ",
"role": "user"
},
{
"content": "As a threshold matter, Section 5 is precisely the sort of statute to which Chevron\nis meant to apply\nAt a mechanical level, Chevron instructs\nto first ask whether the\nstatute is clear\nu\nm\nc\nare inherently ambiguous\nNearly every word\nis rife with ambiguity: What is unfair? Unfair to whom?\nWhat is a method?\nA practice? What is competition? As the Court has noted, the standard is “by necessity\nelusive\nAbsent clarifying language in the statute itself\nthe ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron\nThe sole task of the courts is\nto ensure that\nconstruction is permissible\nThe statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion\nSection 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly\nWhen the Court imposed an overly narrow construction\nCongress amended the statute to overcome that narrowing interpretation\nSection 5 is, thus, a case study in each of the four rationales for Chevron deference\ncongressional intent; agency expertise; concern about\ncourts’ limited political accountability\nand\nseparation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule. | As a threshold matter, Section 5 is precisely the sort of statute to which Chevron
is meant to apply
At a mechanical level, Chevron instructs
to first ask whether the
statute is clear
u
m
c
are inherently ambiguous
Nearly every word
is rife with ambiguity: What is unfair? Unfair to whom?
What is a method?
A practice? What is competition? As the Court has noted, the standard is “by necessity
elusive
Absent clarifying language in the statute itself
the ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron
The sole task of the courts is
to ensure that
construction is permissible
The statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion
Section 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly
When the Court imposed an overly narrow construction
Congress amended the statute to overcome that narrowing interpretation
Section 5 is, thus, a case study in each of the four rationales for Chevron deference
congressional intent; agency expertise; concern about
courts’ limited political accountability
and
separation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference | threshold
precisely the sort of statute to which Chevron
is meant to apply
mechanical
u
m
c
inherently ambiguous
every word
unfair?
whom?
method?
practice?
competition?
Court has noted
by necessity
elusive
Absent clarifying language
statute itself
inherent
sufficient to trigger Chevron
courts
permissible
statutory history
consistently demonstrated a congressional intent to grant the FTC broad discretion
response
courts
too narrowly
amended
overcome
case study in each of the four rationales for Chevron deference
congressional intent; agency expertise; concern about
courts’ limited political accountability
s
o
p
hard to imagine a statute better suited to Chevron deference | 285 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule.",
"role": "system"
},
{
"content": "As a threshold matter, Section 5 is precisely the sort of statute to which Chevron\nis meant to apply\nAt a mechanical level, Chevron instructs\nto first ask whether the\nstatute is clear\nu\nm\nc\nare inherently ambiguous\nNearly every word\nis rife with ambiguity: What is unfair? Unfair to whom?\nWhat is a method?\nA practice? What is competition? As the Court has noted, the standard is “by necessity\nelusive\nAbsent clarifying language in the statute itself\nthe ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron\nThe sole task of the courts is\nto ensure that\nconstruction is permissible\nThe statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion\nSection 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly\nWhen the Court imposed an overly narrow construction\nCongress amended the statute to overcome that narrowing interpretation\nSection 5 is, thus, a case study in each of the four rationales for Chevron deference\ncongressional intent; agency expertise; concern about\ncourts’ limited political accountability\nand\nseparation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference",
"role": "user"
},
{
"content": "threshold\nprecisely the sort of statute to which Chevron\nis meant to apply\nmechanical\nu\nm\nc\ninherently ambiguous\nevery word\nunfair?\nwhom?\nmethod?\npractice?\ncompetition?\nCourt has noted\nby necessity\nelusive\nAbsent clarifying language\nstatute itself\ninherent\nsufficient to trigger Chevron\ncourts\npermissible\nstatutory history\nconsistently demonstrated a congressional intent to grant the FTC broad discretion\nresponse\ncourts\ntoo narrowly\namended\novercome\ncase study in each of the four rationales for Chevron deference\ncongressional intent; agency expertise; concern about\ncourts’ limited political accountability\ns\no\np\nhard to imagine a statute better suited to Chevron deference",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule.
Highlight Ratio: 0.3 | As a threshold matter, Section 5 is precisely the sort of statute to which Chevron
is meant to apply
At a mechanical level, Chevron instructs
to first ask whether the
statute is clear
u
m
c
are inherently ambiguous
Nearly every word
is rife with ambiguity: What is unfair? Unfair to whom?
What is a method?
A practice? What is competition? As the Court has noted, the standard is “by necessity
elusive
Absent clarifying language in the statute itself
the ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron
The sole task of the courts is
to ensure that
construction is permissible
The statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion
Section 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly
When the Court imposed an overly narrow construction
Congress amended the statute to overcome that narrowing interpretation
Section 5 is, thus, a case study in each of the four rationales for Chevron deference
congressional intent; agency expertise; concern about
courts’ limited political accountability
and
separation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference | Section 5 is precisely the sort of statute to which Chevron
apply
every word
rife with ambiguity
Absent clarifying language
ambiguity
is sufficient
The
task of
courts is
to ensure
construction is permissible
statutory history
demonstrated
broad discretion
a case study in
the
rationales for Chevron
intent
expertise
courts’ limited
accountability
and
s
o
p
It is hard to imagine a statute better suited | 286 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: It’s a prototypical Chevron case. Courts would apply deference and uphold the rule.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "As a threshold matter, Section 5 is precisely the sort of statute to which Chevron\nis meant to apply\nAt a mechanical level, Chevron instructs\nto first ask whether the\nstatute is clear\nu\nm\nc\nare inherently ambiguous\nNearly every word\nis rife with ambiguity: What is unfair? Unfair to whom?\nWhat is a method?\nA practice? What is competition? As the Court has noted, the standard is “by necessity\nelusive\nAbsent clarifying language in the statute itself\nthe ambiguity inherent in the language of Section 5 is sufficient to trigger Chevron\nThe sole task of the courts is\nto ensure that\nconstruction is permissible\nThe statutory history has consistently demonstrated a congressional intent to grant the FTC broad discretion\nSection 5 was enacted in response to concerns that the courts had interpreted the antitrust laws too narrowly\nWhen the Court imposed an overly narrow construction\nCongress amended the statute to overcome that narrowing interpretation\nSection 5 is, thus, a case study in each of the four rationales for Chevron deference\ncongressional intent; agency expertise; concern about\ncourts’ limited political accountability\nand\nseparation of powers—all of which urge deference to the FTC’s interpretation of Section 5. It is hard to imagine a statute better suited to Chevron deference",
"role": "user"
},
{
"content": "Section 5 is precisely the sort of statute to which Chevron\napply\nevery word\nrife with ambiguity\nAbsent clarifying language\nambiguity\nis sufficient\nThe\ntask of\ncourts is\nto ensure\nconstruction is permissible\nstatutory history\ndemonstrated\nbroad discretion\na case study in\nthe\nrationales for Chevron\nintent\nexpertise\ncourts’ limited\naccountability\nand\ns\no\np\nIt is hard to imagine a statute better suited",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: First, DETERRENCE---the threat of enforcement deters infractions. | Cases in which the FTC has asserted a broader understanding of Section 5 have generally been resolved in one of two ways: settlement or litigation. Most of the attention paid to the FTC’s expanded use of Section 5 unfair methods of competition claims has focused on high-profile cases that have ultimately settled. When initially bringing a claim, the FTC need not allege anything more than a reason to believe that Section 5 has been violated.147 The FTC need not frame its allegations with any greater specificity; in particular, it need not specify whether it asserts a violation of the traditional antitrust laws (which it can enforce under Section 5) or a standalone Section 5 claim. Rather than limit its options, the FTC typically does not specify a precise legal theory but rather embraces the expansive ambiguity inherent in Section 5’s “unfairness” standard. This approach increases the litigation uncertainty faced by the targets of an FTC investigation, which can be used as leverage by the FTC in securing a favorable settlement.148 This was the pattern used in McWane (discussed below). The FTC also used it in three recent high-profile investigations into high-tech industries: Intel,149 N-Data,150 and Google.151 This approach has also been the basis of the FTC’s privacy and data security jurisprudence, spanning more than one hundred cases.152
While the FTC’s use of Section 5 in high-profile cases has garnered the most attention, its use of Section 5 in lower profile cases, especially those that do not settle, is more revealing. As an administrative agency, a case brought by the FTC is typically heard by an Administrative Law Judge (“ALJ”).153 The FTC prepares and files a complaint, the subject of the investigation files an answer, and both parties submit briefs of their arguments to the ALJ, who will then submit findings of fact and law in an Initial Decision to the Commission.154
A curious thing has happened between the complaint and briefing stages of unfair method of competition cases that the FTC brings before an ALJ. Often, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle, and, typically, the settlement will have been agreed to prior to the filing of the complaint. In those cases that do not settle, the FTC explains in its brief that Section 5 unfair methods of competition claims incorporate Sections 1 and 2 of the Sherman Act.155 | Most
FTC
use of Section 5 u
m
c
claims
focused on high-profile cases that have ultimately settled. When
bringing a claim, the FTC need not allege anything more than a reason to believe
Section 5 has been violated
The FTC need not frame its allegations with any greater specificity
This
increases
litigation uncertainty
which can be used as leverage
in securing a favorable settlement
Often, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle
typically, the settlement will have been agreed to prior to
filing | 287 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: First, DETERRENCE---the threat of enforcement deters infractions.",
"role": "system"
},
{
"content": "Cases in which the FTC has asserted a broader understanding of Section 5 have generally been resolved in one of two ways: settlement or litigation. Most of the attention paid to the FTC’s expanded use of Section 5 unfair methods of competition claims has focused on high-profile cases that have ultimately settled. When initially bringing a claim, the FTC need not allege anything more than a reason to believe that Section 5 has been violated.147 The FTC need not frame its allegations with any greater specificity; in particular, it need not specify whether it asserts a violation of the traditional antitrust laws (which it can enforce under Section 5) or a standalone Section 5 claim. Rather than limit its options, the FTC typically does not specify a precise legal theory but rather embraces the expansive ambiguity inherent in Section 5’s “unfairness” standard. This approach increases the litigation uncertainty faced by the targets of an FTC investigation, which can be used as leverage by the FTC in securing a favorable settlement.148 This was the pattern used in McWane (discussed below). The FTC also used it in three recent high-profile investigations into high-tech industries: Intel,149 N-Data,150 and Google.151 This approach has also been the basis of the FTC’s privacy and data security jurisprudence, spanning more than one hundred cases.152 \nWhile the FTC’s use of Section 5 in high-profile cases has garnered the most attention, its use of Section 5 in lower profile cases, especially those that do not settle, is more revealing. As an administrative agency, a case brought by the FTC is typically heard by an Administrative Law Judge (“ALJ”).153 The FTC prepares and files a complaint, the subject of the investigation files an answer, and both parties submit briefs of their arguments to the ALJ, who will then submit findings of fact and law in an Initial Decision to the Commission.154 \nA curious thing has happened between the complaint and briefing stages of unfair method of competition cases that the FTC brings before an ALJ. Often, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle, and, typically, the settlement will have been agreed to prior to the filing of the complaint. In those cases that do not settle, the FTC explains in its brief that Section 5 unfair methods of competition claims incorporate Sections 1 and 2 of the Sherman Act.155 ",
"role": "user"
},
{
"content": "Most\nFTC\nuse of Section 5 u\nm\nc\nclaims\nfocused on high-profile cases that have ultimately settled. When\nbringing a claim, the FTC need not allege anything more than a reason to believe\nSection 5 has been violated\nThe FTC need not frame its allegations with any greater specificity\nThis\nincreases\nlitigation uncertainty\nwhich can be used as leverage\nin securing a favorable settlement\nOften, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle\ntypically, the settlement will have been agreed to prior to\nfiling",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: First, DETERRENCE---the threat of enforcement deters infractions. | Most
FTC
use of Section 5 u
m
c
claims
focused on high-profile cases that have ultimately settled. When
bringing a claim, the FTC need not allege anything more than a reason to believe
Section 5 has been violated
The FTC need not frame its allegations with any greater specificity
This
increases
litigation uncertainty
which can be used as leverage
in securing a favorable settlement
Often, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle
typically, the settlement will have been agreed to prior to
filing | Most
FTC
u
m
c
focus
high-profile cases
settled
reason to believe
specificity
uncertainty
leverage
settlement
only Section 5
vast majority of cases, this is sufficient to spur the target of the investigation to settle
prior to
filing | 288 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: First, DETERRENCE---the threat of enforcement deters infractions.",
"role": "system"
},
{
"content": "Most\nFTC\nuse of Section 5 u\nm\nc\nclaims\nfocused on high-profile cases that have ultimately settled. When\nbringing a claim, the FTC need not allege anything more than a reason to believe\nSection 5 has been violated\nThe FTC need not frame its allegations with any greater specificity\nThis\nincreases\nlitigation uncertainty\nwhich can be used as leverage\nin securing a favorable settlement\nOften, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle\ntypically, the settlement will have been agreed to prior to\nfiling",
"role": "user"
},
{
"content": "Most\nFTC\nu\nm\nc\nfocus\nhigh-profile cases\nsettled\nreason to believe\nspecificity\nuncertainty\nleverage\nsettlement\nonly Section 5\nvast majority of cases, this is sufficient to spur the target of the investigation to settle\nprior to\nfiling",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: First, DETERRENCE---the threat of enforcement deters infractions.
Highlight Ratio: 0.4 | Most
FTC
use of Section 5 u
m
c
claims
focused on high-profile cases that have ultimately settled. When
bringing a claim, the FTC need not allege anything more than a reason to believe
Section 5 has been violated
The FTC need not frame its allegations with any greater specificity
This
increases
litigation uncertainty
which can be used as leverage
in securing a favorable settlement
Often, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle
typically, the settlement will have been agreed to prior to
filing | FTC
Section 5 u
m
c
claims
high-profile cases
have ultimately settled
litigation uncertainty
can be
leverage
in securing a
settlement
the complaint
cite only Section 5
In the vast majority of cases, this is sufficient to
settle
prior to
filing | 289 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: First, DETERRENCE---the threat of enforcement deters infractions.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "Most\nFTC\nuse of Section 5 u\nm\nc\nclaims\nfocused on high-profile cases that have ultimately settled. When\nbringing a claim, the FTC need not allege anything more than a reason to believe\nSection 5 has been violated\nThe FTC need not frame its allegations with any greater specificity\nThis\nincreases\nlitigation uncertainty\nwhich can be used as leverage\nin securing a favorable settlement\nOften, the complaint will cite only Section 5 as the legal basis for the complaint. In the vast majority of cases, this is sufficient to spur the target of the investigation to settle\ntypically, the settlement will have been agreed to prior to\nfiling",
"role": "user"
},
{
"content": "FTC\nSection 5 u\nm\nc\nclaims\nhigh-profile cases\nhave ultimately settled\nlitigation uncertainty\ncan be\nleverage\nin securing a\nsettlement\nthe complaint\ncite only Section 5\nIn the vast majority of cases, this is sufficient to\nsettle\nprior to\nfiling",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: That solves even without a ruling | Even if the FTC cannot persuade the courts that the federal antitrust laws should prohibit reverse-payment settlements, it can reach beyond traditional federal antitrust law and rely instead on the unfair competition prong of Section 5 of the FTC Act to challenge these arrangements. Although most commentators agree that Section 5 is broader than the Sherman Act, the FTC rarely has invoked it as the stand-alone basis for a cause of action to reach conduct that would be legal under the antitrust laws. The FTC did test Section 5’s boundaries in the early 1980s, but mostly was rebuffed and, until very recently, has shied away from basing its actions exclusively on Section 5. Recent statements and speeches by some FTC commissioners, however, have emphasized that the agency, as a matter of policy, is determined to seek opportunities to bring pure Section 5 cases in the future. Indeed, in late 2009, the FTC exercised this policy goal when it filed a massive antitrust action against Intel that was expressly based upon a substantive Section 5 claim. The law governing Section 5 claims is sparse and, not surprisingly, no court has yet ruled on the legality of reverse-payment settlements under Section 5 alone. Even absent such a ruling, companies considering a reverse-payment settlement should proceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of its ultimate merit. | Even if the FTC cannot persuade the courts that
federal antitrust laws should prohibit
it can reach beyond traditional
antitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A
most commentators agree that Section 5 is broader than
Sherman
the FTC rarely
invoked it as the stand-alone basis for a cause of action to reach conduct that would be legal
no court has yet ruled on
legality
under Section 5 alone. Even absent such a ruling, companies considering
proceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of
ultimate merit | 290 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: That solves even without a ruling",
"role": "system"
},
{
"content": "Even if the FTC cannot persuade the courts that the federal antitrust laws should prohibit reverse-payment settlements, it can reach beyond traditional federal antitrust law and rely instead on the unfair competition prong of Section 5 of the FTC Act to challenge these arrangements. Although most commentators agree that Section 5 is broader than the Sherman Act, the FTC rarely has invoked it as the stand-alone basis for a cause of action to reach conduct that would be legal under the antitrust laws. The FTC did test Section 5’s boundaries in the early 1980s, but mostly was rebuffed and, until very recently, has shied away from basing its actions exclusively on Section 5. Recent statements and speeches by some FTC commissioners, however, have emphasized that the agency, as a matter of policy, is determined to seek opportunities to bring pure Section 5 cases in the future. Indeed, in late 2009, the FTC exercised this policy goal when it filed a massive antitrust action against Intel that was expressly based upon a substantive Section 5 claim. The law governing Section 5 claims is sparse and, not surprisingly, no court has yet ruled on the legality of reverse-payment settlements under Section 5 alone. Even absent such a ruling, companies considering a reverse-payment settlement should proceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of its ultimate merit.",
"role": "user"
},
{
"content": "Even if the FTC cannot persuade the courts that\nfederal antitrust laws should prohibit\nit can reach beyond traditional\nantitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A\nmost commentators agree that Section 5 is broader than\nSherman\nthe FTC rarely\ninvoked it as the stand-alone basis for a cause of action to reach conduct that would be legal\nno court has yet ruled on\nlegality\nunder Section 5 alone. Even absent such a ruling, companies considering\nproceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of\nultimate merit",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: That solves even without a ruling | Even if the FTC cannot persuade the courts that
federal antitrust laws should prohibit
it can reach beyond traditional
antitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A
most commentators agree that Section 5 is broader than
Sherman
the FTC rarely
invoked it as the stand-alone basis for a cause of action to reach conduct that would be legal
no court has yet ruled on
legality
under Section 5 alone. Even absent such a ruling, companies considering
proceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of
ultimate merit | persuade the courts
prohibit
beyond traditional
antitrust law
unfair competition prong of Section 5 of the FTC A
most commentators agree
broader
Sherman
rarely
invoked it
stand-alone basis
would
legal
ruled
legality
Section 5 alone
absent
considering
proceed with caution
FTC investigation
significantly burdensome
expensive regardless
ultimate merit | 291 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: That solves even without a ruling",
"role": "system"
},
{
"content": "Even if the FTC cannot persuade the courts that\nfederal antitrust laws should prohibit\nit can reach beyond traditional\nantitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A\nmost commentators agree that Section 5 is broader than\nSherman\nthe FTC rarely\ninvoked it as the stand-alone basis for a cause of action to reach conduct that would be legal\nno court has yet ruled on\nlegality\nunder Section 5 alone. Even absent such a ruling, companies considering\nproceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of\nultimate merit",
"role": "user"
},
{
"content": "persuade the courts\nprohibit\nbeyond traditional\nantitrust law\nunfair competition prong of Section 5 of the FTC A\nmost commentators agree\nbroader\nSherman\nrarely\ninvoked it\nstand-alone basis\nwould\nlegal\nruled\nlegality\nSection 5 alone\nabsent\nconsidering\nproceed with caution\nFTC investigation\nsignificantly burdensome\nexpensive regardless\nultimate merit",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: That solves even without a ruling
Highlight Ratio: 0.4 | Even if the FTC cannot persuade the courts that
federal antitrust laws should prohibit
it can reach beyond traditional
antitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A
most commentators agree that Section 5 is broader than
Sherman
the FTC rarely
invoked it as the stand-alone basis for a cause of action to reach conduct that would be legal
no court has yet ruled on
legality
under Section 5 alone. Even absent such a ruling, companies considering
proceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of
ultimate merit | Even if
FTC cannot persuade
courts
antitrust
should prohibit
it can
rely
on
Section 5
no court
yet ruled on
Section 5 alone. Even absent such
companies
proceed with caution, as
FTC investigation can be
nsome and expensive regardless of
ultimate merit | 292 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: That solves even without a ruling\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "Even if the FTC cannot persuade the courts that\nfederal antitrust laws should prohibit\nit can reach beyond traditional\nantitrust law and rely instead on the unfair competition prong of Section 5 of the FTC A\nmost commentators agree that Section 5 is broader than\nSherman\nthe FTC rarely\ninvoked it as the stand-alone basis for a cause of action to reach conduct that would be legal\nno court has yet ruled on\nlegality\nunder Section 5 alone. Even absent such a ruling, companies considering\nproceed with caution, as an FTC investigation can be significantly burdensome and expensive regardless of\nultimate merit",
"role": "user"
},
{
"content": "Even if\nFTC cannot persuade\ncourts\nantitrust\nshould prohibit\nit can\nrely\non\nSection 5\nno court\nyet ruled on\nSection 5 alone. Even absent such\ncompanies\nproceed with caution, as\nFTC investigation can be\nnsome and expensive regardless of\nultimate merit",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement. | Relatedly, for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say, (1) here are our standards for Section 5, and (2) here’s why we think this firm violated them, that strikes me as much stronger than something along the lines of, “we didn’t like what this firm was doing so we trotted out section 5 to deal with it.” The former position seems much more likely to get at least Skidmore deference, the latter position, judicial mockery.
Externally, and I’m sure others have said this, for those subject to Section 5 (and by this I suppose I mean just about every business other than common carriers) a policy statement would obviously eliminate some uncertainty, particularly if the agency repudiated some of the wilder visions of what section 5 covers. For example, I think the FTC could and should walk away from: | for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say
here are our standards for Section 5, and
here’s why we think this firm violated them, that strikes
much stronger than
we didn’t like
this
so we trotted out section 5
The former
seems much more likely to get
deference, the latter
mockery
for those subject to Section 5
a policy
would
eliminate
uncertainty
if the agency repudiated some of the wilder visions of what section 5 covers | 293 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement.",
"role": "system"
},
{
"content": "Relatedly, for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say, (1) here are our standards for Section 5, and (2) here’s why we think this firm violated them, that strikes me as much stronger than something along the lines of, “we didn’t like what this firm was doing so we trotted out section 5 to deal with it.” The former position seems much more likely to get at least Skidmore deference, the latter position, judicial mockery.\nExternally, and I’m sure others have said this, for those subject to Section 5 (and by this I suppose I mean just about every business other than common carriers) a policy statement would obviously eliminate some uncertainty, particularly if the agency repudiated some of the wilder visions of what section 5 covers. For example, I think the FTC could and should walk away from:",
"role": "user"
},
{
"content": "for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say\nhere are our standards for Section 5, and\nhere’s why we think this firm violated them, that strikes\nmuch stronger than\nwe didn’t like\nthis\nso we trotted out section 5\nThe former\nseems much more likely to get\ndeference, the latter\nmockery\nfor those subject to Section 5\na policy\nwould\neliminate\nuncertainty\nif the agency repudiated some of the wilder visions of what section 5 covers",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement. | for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say
here are our standards for Section 5, and
here’s why we think this firm violated them, that strikes
much stronger than
we didn’t like
this
so we trotted out section 5
The former
seems much more likely to get
deference, the latter
mockery
for those subject to Section 5
a policy
would
eliminate
uncertainty
if the agency repudiated some of the wilder visions of what section 5 covers | useful
litigation
here are our standards for Section 5, and
here’s why we think this firm violated them
that s
much stronger
we didn’t like
this
so we trotted out section 5
former
much more likely to get
deference
latter
mockery
subject
policy
eliminate
uncertainty
wilder | 294 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement.",
"role": "system"
},
{
"content": "for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say\nhere are our standards for Section 5, and\nhere’s why we think this firm violated them, that strikes\nmuch stronger than\nwe didn’t like\nthis\nso we trotted out section 5\nThe former\nseems much more likely to get\ndeference, the latter\nmockery\nfor those subject to Section 5\na policy\nwould\neliminate\nuncertainty\nif the agency repudiated some of the wilder visions of what section 5 covers",
"role": "user"
},
{
"content": "useful\nlitigation\nhere are our standards for Section 5, and\nhere’s why we think this firm violated them\nthat s\nmuch stronger\nwe didn’t like\nthis\nso we trotted out section 5\nformer\nmuch more likely to get\ndeference\nlatter\nmockery\nsubject\npolicy\neliminate\nuncertainty\nwilder",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement.
Highlight Ratio: 0.6 | for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say
here are our standards for Section 5, and
here’s why we think this firm violated them, that strikes
much stronger than
we didn’t like
this
so we trotted out section 5
The former
seems much more likely to get
deference, the latter
mockery
for those subject to Section 5
a policy
would
eliminate
uncertainty
if the agency repudiated some of the wilder visions of what section 5 covers | a policy
could be
useful
for litigation. If the agency
stand up
and say
here are our standards for Section 5, and
here’s why
this firm violated them, that s
much stronger than
we didn’t like
this
so we trotted out section 5
The former
seems
likely to get
deference, the latter
mockery
policy
would
eliminate
uncertainty | 295 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Deference is way more likely if there’s a rule rather than just ad hoc enforcement.\n\nHighlight Ratio: 0.6",
"role": "system"
},
{
"content": "for the agency’s work, a policy statement could be a useful thing for litigation. If the agency can stand up in appellate court and say\nhere are our standards for Section 5, and\nhere’s why we think this firm violated them, that strikes\nmuch stronger than\nwe didn’t like\nthis\nso we trotted out section 5\nThe former\nseems much more likely to get\ndeference, the latter\nmockery\nfor those subject to Section 5\na policy\nwould\neliminate\nuncertainty\nif the agency repudiated some of the wilder visions of what section 5 covers",
"role": "user"
},
{
"content": "a policy\ncould be\nuseful\nfor litigation. If the agency\nstand up\nand say\nhere are our standards for Section 5, and\nhere’s why\nthis firm violated them, that s\nmuch stronger than\nwe didn’t like\nthis\nso we trotted out section 5\nThe former\nseems\nlikely to get\ndeference, the latter\nmockery\npolicy\nwould\neliminate\nuncertainty",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Jones and Kovacic is about FTCA Section 5 enforcement actions, NOT about rulemaking. | One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution, through its own administrative process, of cases based on Section 5 of the FTC Act and its prohibition of “unfair methods of competition.”93 This section allows the FTC94 to tackle not only anticompetitive practices prohibited by the other antitrust statutes but also conduct constituting incipient violations of those statutes or behavior that exceeds their reach. The latter is possible where the conduct does not infringe the letter of the antitrust laws but contradicts their basic spirit or public policy.95
There is no doubt therefore that Section 5 was designed as an expansion joint in the U.S. antitrust system. It seems unlikely to us, nonetheless, that a majority of FTC’s current members will be minded to use it in this way. Further, even if they were to be, the reality is that such an application may encounter difficulties. Since its creation in 1914, the FTC has never prevailed before the Supreme Court in any case challenging dominant firm misconduct, whether premised on Section 2 of the Sherman Act or purely on Section 5 of the FTC Act.96 The last FTC success in federal court in a case predicated solely on Section 5 occurred in the late 1960s.97
The FTC’s record of limited success with Section 5 has not been for want of trying. In the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5, a foundation to develop “competition policy in its broadest sense.”98 The agency’s Section 5 agenda yielded some successes,99 but also a large number of litigation failures involving cases to address subtle forms of coordination in oligopolies, to impose new obligations on dominant firms, and to dissolve shared monopolies.100 The agency’s program elicited powerful legislative backlash from a Congress that once supported FTC’s trailblazing initiatives but turned against it as the Commission’s efforts to obtain dramatic structural remedies unfolded.101 | One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution
based on Section 5 of the FTC Act
In the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5 | 296 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Jones and Kovacic is about FTCA Section 5 enforcement actions, NOT about rulemaking.",
"role": "system"
},
{
"content": "One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution, through its own administrative process, of cases based on Section 5 of the FTC Act and its prohibition of “unfair methods of competition.”93 This section allows the FTC94 to tackle not only anticompetitive practices prohibited by the other antitrust statutes but also conduct constituting incipient violations of those statutes or behavior that exceeds their reach. The latter is possible where the conduct does not infringe the letter of the antitrust laws but contradicts their basic spirit or public policy.95\nThere is no doubt therefore that Section 5 was designed as an expansion joint in the U.S. antitrust system. It seems unlikely to us, nonetheless, that a majority of FTC’s current members will be minded to use it in this way. Further, even if they were to be, the reality is that such an application may encounter difficulties. Since its creation in 1914, the FTC has never prevailed before the Supreme Court in any case challenging dominant firm misconduct, whether premised on Section 2 of the Sherman Act or purely on Section 5 of the FTC Act.96 The last FTC success in federal court in a case predicated solely on Section 5 occurred in the late 1960s.97\nThe FTC’s record of limited success with Section 5 has not been for want of trying. In the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5, a foundation to develop “competition policy in its broadest sense.”98 The agency’s Section 5 agenda yielded some successes,99 but also a large number of litigation failures involving cases to address subtle forms of coordination in oligopolies, to impose new obligations on dominant firms, and to dissolve shared monopolies.100 The agency’s program elicited powerful legislative backlash from a Congress that once supported FTC’s trailblazing initiatives but turned against it as the Commission’s efforts to obtain dramatic structural remedies unfolded.101",
"role": "user"
},
{
"content": "One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution\nbased on Section 5 of the FTC Act\nIn the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Jones and Kovacic is about FTCA Section 5 enforcement actions, NOT about rulemaking.
Highlight Ratio: 0.4 | One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution
based on Section 5 of the FTC Act
In the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5 | One
solution
is for the FTC to rely
on
prosecution
based on Section 5
the FTC undertook
ambitious
enforcement of claims predicated on
Section 5 | 298 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Jones and Kovacic is about FTCA Section 5 enforcement actions, NOT about rulemaking.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "One possible solution to rigidities that have developed in Sherman Act jurisprudence is for the FTC to rely more heavily on the prosecution\nbased on Section 5 of the FTC Act\nIn the 1970s, the FTC undertook an ambitious program to make the enforcement of claims predicated on the distinctive reach of Section 5",
"role": "user"
},
{
"content": "One\nsolution\nis for the FTC to rely\non\nprosecution\nbased on Section 5\nthe FTC undertook\nambitious\nenforcement of claims predicated on\nSection 5",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Past decisions prove it would be covered. | Had Chevron deference existed in the 1970s, the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation. Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity. Granted, the DOJ was acting as an antitrust enforcer with the same goal, but the Supreme Court regularly recognized section 5's separate and ambiguous scope.96 | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | 299 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Past decisions prove it would be covered.",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s, the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation. Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity. Granted, the DOJ was acting as an antitrust enforcer with the same goal, but the Supreme Court regularly recognized section 5's separate and ambiguous scope.96 ",
"role": "user"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Past decisions prove it would be covered. | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | 70s
FTC
clear
that time
exclusive rulemaking authority over a statute with a unique reach and interpretation
initial requirement
delegation through ambiguity
regularly recognized
separate
ambiguous scope | 300 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Past decisions prove it would be covered.",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "user"
},
{
"content": "70s\nFTC\nclear\nthat time\nexclusive rulemaking authority over a statute with a unique reach and interpretation\ninitial requirement\ndelegation through ambiguity\nregularly recognized\nseparate\nambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: Past decisions prove it would be covered.
Highlight Ratio: 0.8 | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | Had Chevron
existed in the
70s
the case for
deference to
FTC would have been clear
the FTC exercised exclusive rulemaking authority over a statute with
reach and interpretation
the initial requirement for
deference was present: delegation through ambiguity
the
Court regularly recognized section 5's separate and ambiguous scope | 301 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: Past decisions prove it would be covered.\n\nHighlight Ratio: 0.8",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "user"
},
{
"content": "Had Chevron\nexisted in the\n70s\nthe case for\ndeference to\nFTC would have been clear\nthe FTC exercised exclusive rulemaking authority over a statute with\nreach and interpretation\nthe initial requirement for\ndeference was present: delegation through ambiguity\nthe\nCourt regularly recognized section 5's separate and ambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: If it had, there’d have been a clear case for deference. | Had Chevron deference existed in the 1970s, the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation. Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity. Granted, the DOJ was acting as an antitrust enforcer with the same goal, but the Supreme Court regularly recognized section 5's separate and ambiguous scope.96 | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | 302 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: If it had, there’d have been a clear case for deference.",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s, the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation. Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity. Granted, the DOJ was acting as an antitrust enforcer with the same goal, but the Supreme Court regularly recognized section 5's separate and ambiguous scope.96 ",
"role": "user"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: If it had, there’d have been a clear case for deference. | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | 70s
FTC
clear
that time
exclusive rulemaking authority over a statute with a unique reach and interpretation
initial requirement
delegation through ambiguity
regularly recognized
separate
ambiguous scope | 303 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: If it had, there’d have been a clear case for deference.",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "user"
},
{
"content": "70s\nFTC\nclear\nthat time\nexclusive rulemaking authority over a statute with a unique reach and interpretation\ninitial requirement\ndelegation through ambiguity\nregularly recognized\nseparate\nambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: If it had, there’d have been a clear case for deference.
Highlight Ratio: 0.8 | Had Chevron deference existed in the 1970s
the case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation
Likewise, the initial requirement for Chevron deference was present: delegation through ambiguity
the Supreme Court regularly recognized section 5's separate and ambiguous scope | Had Chevron
existed in the
70s
the case for
deference to
FTC would have been clear
the FTC exercised exclusive rulemaking authority over a statute with
reach and interpretation
the initial requirement for
deference was present: delegation through ambiguity
the
Court regularly recognized section 5's separate and ambiguous scope | 304 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: If it had, there’d have been a clear case for deference.\n\nHighlight Ratio: 0.8",
"role": "system"
},
{
"content": "Had Chevron deference existed in the 1970s\nthe case for granting deference to the FTC would have been clear. At that time, the FTC exercised exclusive rulemaking authority over a statute with a unique reach and interpretation\nLikewise, the initial requirement for Chevron deference was present: delegation through ambiguity\nthe Supreme Court regularly recognized section 5's separate and ambiguous scope",
"role": "user"
},
{
"content": "Had Chevron\nexisted in the\n70s\nthe case for\ndeference to\nFTC would have been clear\nthe FTC exercised exclusive rulemaking authority over a statute with\nreach and interpretation\nthe initial requirement for\ndeference was present: delegation through ambiguity\nthe\nCourt regularly recognized section 5's separate and ambiguous scope",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations. | C. Indiana Federation of Dentists Supports Application of Chevron
Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review of FTC legal determinations under Section 5, according some (limited) deference to the FTC.219 It also bears special note as it is the most recent Supreme Court opinion considering the application of Section 5’s prohibition against unfair methods of competition—the only opinion to do so since Chevron.220
The most cited passage from Indiana Federation of Dentists explains that:
The legal issues presented—that is, the identification of governing legal standards and their application to the facts found—are, by contrast, for the courts to resolve, although even in considering such issues the courts are to give some deference to the [FTC’s] informed judgment that a particular commercial practice is to be condemned as “unfair.”221
This language has been cited as requiring do novo review of all legal questions, including the legal meaning of Section 5.222 Dan Crane has called this an “odd standard,”223 noting that ordinarily “this is technically a question of Chevron deference, although the courts have not articulated it that way in the antitrust space.”224 Indeed, it seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led antitrust commentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms.”225 But this is an over-reading of Indiana Federation of Dentists. Indeed, the case can instead be read as entirely in line with Chevron.
First, it is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided, and the FTC was arguing for an uncontroversial interpretation of Section 5 as applying Section 1 of the Sherman Act.226 In other words, the FTC had never structured its case to seek deference, and it had no need to argue for any deference before the Court. Given the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.
Moreover, it took several years for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation. Over the next several years, the circuits regularly cited Indiana Federation of Dentists to explain the standard of review for an agency’s interpretation of its organic statutes.227 Importantly, these cases recognized that there was some confusion as to the changing standard of review,228 framed their analysis in terms of Skidmore (the precursor to Chevron in this line of cases),229 and largely reached Chevron-like conclusions, despite Indiana Federation of Dentists’s suggestion of a lower level of deference.230 Perhaps most importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC.
Indeed, a close reading of Indiana Federation of Dentists finds that it accords with Chevron. The continuation of the paragraph quoted above goes on to explain that:
The standard of “unfairness” under the [FTCA] is, by necessity, an elusive one, encompassing not only practices that violate the Sherman Act and the other antitrust laws, but also practices that the [FTC] determines are against public policy for other reasons. Once the [FTC] has chosen a particular legal rationale for holding a practice to be unfair, however, familiar principles of administrative law dictate that its decision must stand or fall on that basis, and a reviewing court may not consider other reasons why the practice might be deemed unfair. In the case now before us, the sole basis of the FTC’s finding of an unfair method of competition was [its] conclusion that the [alleged conduct] was an unreasonable and conspiratorial restraint of trade in violation of § 1 of the Sherman Act. Accordingly, the legal question before us is whether the [FTC’s] factual findings, if supported by evidence, make out a violation of Sherman Act § 1.231
This language alters the paragraph’s initial proposition that the legal issues are for determination by the courts. Rather, the Court recognizes that Section 5 is inherently ambiguous. It is, therefore, up to the FTC to choose the legal standard under which that conduct will be reviewed: “[A] reviewing court may not consider other reasons why the practice might be deemed unfair.”
This is precisely the standard established by Chevron: First, the courts determine whether the statute is ambiguous, and, if it is not, the court’s reading of the statute is binding; but if it is ambiguous, the court defers to the agency’s construction.232 Part of why Chevron is a difficult test is that both parts of this analysis do, in fact, present legal questions for the court. The first step is purely legal, as the court determines on its own whether the statute is ambiguous. Then, at step two, the legal question is whether the agency correctly applied the facts to its declared legal standard—as the Court recognized in Indiana Federation of Dentists, “the legal question before us is whether the FTC’s factual findings make out a violation of Sherman Act § 1.”233 Thus, the opening, oft-quoted first sentence of the paragraph234 is correct and in accord with Chevron: The legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC.
The most likely reason that Indiana Federation of Dentists is viewed as the standard of review for the FTC’s interpretation of Section 5 is because the FTC has not sought greater deference. This is in part because, where the FTC couches enforcement of Section 5 in the antitrust laws, it can safely rely on judicially- crafted understandings of the antitrust laws without any need to seek deference. Thus, in Schering-Plough, where the FTC’s finding was based on Section 1 of the Sherman Act, the FTC’s brief recounted Indiana Federation of Dentists as requiring de novo review of its legal determinations—a standard that was then used by the Eleventh Circuit in its opinion.235 But it is also surely in part because the FTC has been reluctant to advance a more deferential standard (shell-shocked as it is from pre-Chevron losses)236 and has failed to recognize the current agency- deferential state of administrative law. | Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review
it seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led
commentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms
this is an over-reading
Indeed, the case can instead be read as entirely in line with Chevron.
it is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided
the FTC was arguing for an uncontroversial interpretation of Section 5
the FTC had never structured its case to seek deference, and it had no need to argue for any deference
Given the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.
Moreover, it took
years for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation
most importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC
the Court recognizes that Section 5 is inherently ambiguous. It is, therefore
up to the FTC to choose the legal standard under which
conduct will be reviewed
This is precisely the standard established by Chevron
The legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC
because the FTC has been reluctant to advance a more deferential standard
shell-shocked as it is from pre-Chevron losses
and has failed to recognize the current agency- deferential state of administrative law | 305 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations.",
"role": "system"
},
{
"content": "C. Indiana Federation of Dentists Supports Application of Chevron \nIndiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review of FTC legal determinations under Section 5, according some (limited) deference to the FTC.219 It also bears special note as it is the most recent Supreme Court opinion considering the application of Section 5’s prohibition against unfair methods of competition—the only opinion to do so since Chevron.220 \nThe most cited passage from Indiana Federation of Dentists explains that: \nThe legal issues presented—that is, the identification of governing legal standards and their application to the facts found—are, by contrast, for the courts to resolve, although even in considering such issues the courts are to give some deference to the [FTC’s] informed judgment that a particular commercial practice is to be condemned as “unfair.”221 \nThis language has been cited as requiring do novo review of all legal questions, including the legal meaning of Section 5.222 Dan Crane has called this an “odd standard,”223 noting that ordinarily “this is technically a question of Chevron deference, although the courts have not articulated it that way in the antitrust space.”224 Indeed, it seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led antitrust commentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms.”225 But this is an over-reading of Indiana Federation of Dentists. Indeed, the case can instead be read as entirely in line with Chevron. \nFirst, it is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided, and the FTC was arguing for an uncontroversial interpretation of Section 5 as applying Section 1 of the Sherman Act.226 In other words, the FTC had never structured its case to seek deference, and it had no need to argue for any deference before the Court. Given the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron. \nMoreover, it took several years for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation. Over the next several years, the circuits regularly cited Indiana Federation of Dentists to explain the standard of review for an agency’s interpretation of its organic statutes.227 Importantly, these cases recognized that there was some confusion as to the changing standard of review,228 framed their analysis in terms of Skidmore (the precursor to Chevron in this line of cases),229 and largely reached Chevron-like conclusions, despite Indiana Federation of Dentists’s suggestion of a lower level of deference.230 Perhaps most importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC. \nIndeed, a close reading of Indiana Federation of Dentists finds that it accords with Chevron. The continuation of the paragraph quoted above goes on to explain that: \nThe standard of “unfairness” under the [FTCA] is, by necessity, an elusive one, encompassing not only practices that violate the Sherman Act and the other antitrust laws, but also practices that the [FTC] determines are against public policy for other reasons. Once the [FTC] has chosen a particular legal rationale for holding a practice to be unfair, however, familiar principles of administrative law dictate that its decision must stand or fall on that basis, and a reviewing court may not consider other reasons why the practice might be deemed unfair. In the case now before us, the sole basis of the FTC’s finding of an unfair method of competition was [its] conclusion that the [alleged conduct] was an unreasonable and conspiratorial restraint of trade in violation of § 1 of the Sherman Act. Accordingly, the legal question before us is whether the [FTC’s] factual findings, if supported by evidence, make out a violation of Sherman Act § 1.231 \nThis language alters the paragraph’s initial proposition that the legal issues are for determination by the courts. Rather, the Court recognizes that Section 5 is inherently ambiguous. It is, therefore, up to the FTC to choose the legal standard under which that conduct will be reviewed: “[A] reviewing court may not consider other reasons why the practice might be deemed unfair.” \nThis is precisely the standard established by Chevron: First, the courts determine whether the statute is ambiguous, and, if it is not, the court’s reading of the statute is binding; but if it is ambiguous, the court defers to the agency’s construction.232 Part of why Chevron is a difficult test is that both parts of this analysis do, in fact, present legal questions for the court. The first step is purely legal, as the court determines on its own whether the statute is ambiguous. Then, at step two, the legal question is whether the agency correctly applied the facts to its declared legal standard—as the Court recognized in Indiana Federation of Dentists, “the legal question before us is whether the FTC’s factual findings make out a violation of Sherman Act § 1.”233 Thus, the opening, oft-quoted first sentence of the paragraph234 is correct and in accord with Chevron: The legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC. \nThe most likely reason that Indiana Federation of Dentists is viewed as the standard of review for the FTC’s interpretation of Section 5 is because the FTC has not sought greater deference. This is in part because, where the FTC couches enforcement of Section 5 in the antitrust laws, it can safely rely on judicially- crafted understandings of the antitrust laws without any need to seek deference. Thus, in Schering-Plough, where the FTC’s finding was based on Section 1 of the Sherman Act, the FTC’s brief recounted Indiana Federation of Dentists as requiring de novo review of its legal determinations—a standard that was then used by the Eleventh Circuit in its opinion.235 But it is also surely in part because the FTC has been reluctant to advance a more deferential standard (shell-shocked as it is from pre-Chevron losses)236 and has failed to recognize the current agency- deferential state of administrative law. ",
"role": "user"
},
{
"content": "Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review\nit seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led\ncommentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms\nthis is an over-reading\nIndeed, the case can instead be read as entirely in line with Chevron.\nit is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided\nthe FTC was arguing for an uncontroversial interpretation of Section 5\nthe FTC had never structured its case to seek deference, and it had no need to argue for any deference\nGiven the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.\nMoreover, it took\nyears for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation\nmost importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC\nthe Court recognizes that Section 5 is inherently ambiguous. It is, therefore\nup to the FTC to choose the legal standard under which\nconduct will be reviewed\nThis is precisely the standard established by Chevron\nThe legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC\nbecause the FTC has been reluctant to advance a more deferential standard\nshell-shocked as it is from pre-Chevron losses\nand has failed to recognize the current agency- deferential state of administrative law",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations. | Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review
it seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led
commentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms
this is an over-reading
Indeed, the case can instead be read as entirely in line with Chevron.
it is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided
the FTC was arguing for an uncontroversial interpretation of Section 5
the FTC had never structured its case to seek deference, and it had no need to argue for any deference
Given the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.
Moreover, it took
years for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation
most importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC
the Court recognizes that Section 5 is inherently ambiguous. It is, therefore
up to the FTC to choose the legal standard under which
conduct will be reviewed
This is precisely the standard established by Chevron
The legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC
because the FTC has been reluctant to advance a more deferential standard
shell-shocked as it is from pre-Chevron losses
and has failed to recognize the current agency- deferential state of administrative law | special discussion
de novo review
mention Chevron
cannot explain judicial posture in the antitrust arena in Chevron terms
over-reading
instead
entirely in line with Chevron
argued before Chevron was decided
uncontroversial interpretation of Section 5
seek
had no need to argue for any deference
more surprising
cited
years
importance
filter its way into judicial review
today
Chevron
every regulatory agency other than the FTC
recognizes
inherently ambiguous
up to the FTC to choose the legal standard under which
conduct will be reviewed
precisely the standard established by Chevron
according
legal standard prescribed by the FTC
advance
shell-shocked
pre-Chevron losses
failed to recognize the current agency- deferential state of administrative law | 306 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations.",
"role": "system"
},
{
"content": "Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review\nit seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led\ncommentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms\nthis is an over-reading\nIndeed, the case can instead be read as entirely in line with Chevron.\nit is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided\nthe FTC was arguing for an uncontroversial interpretation of Section 5\nthe FTC had never structured its case to seek deference, and it had no need to argue for any deference\nGiven the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.\nMoreover, it took\nyears for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation\nmost importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC\nthe Court recognizes that Section 5 is inherently ambiguous. It is, therefore\nup to the FTC to choose the legal standard under which\nconduct will be reviewed\nThis is precisely the standard established by Chevron\nThe legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC\nbecause the FTC has been reluctant to advance a more deferential standard\nshell-shocked as it is from pre-Chevron losses\nand has failed to recognize the current agency- deferential state of administrative law",
"role": "user"
},
{
"content": "special discussion\nde novo review\nmention Chevron\ncannot explain judicial posture in the antitrust arena in Chevron terms\nover-reading\ninstead\nentirely in line with Chevron\nargued before Chevron was decided\nuncontroversial interpretation of Section 5\nseek\nhad no need to argue for any deference\nmore surprising\ncited\nyears\nimportance\nfilter its way into judicial review\ntoday\nChevron\nevery regulatory agency other than the FTC\nrecognizes\ninherently ambiguous\nup to the FTC to choose the legal standard under which\nconduct will be reviewed\nprecisely the standard established by Chevron\naccording\nlegal standard prescribed by the FTC\nadvance\nshell-shocked\npre-Chevron losses\nfailed to recognize the current agency- deferential state of administrative law",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations.
Highlight Ratio: 0.4 | Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review
it seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led
commentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms
this is an over-reading
Indeed, the case can instead be read as entirely in line with Chevron.
it is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided
the FTC was arguing for an uncontroversial interpretation of Section 5
the FTC had never structured its case to seek deference, and it had no need to argue for any deference
Given the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.
Moreover, it took
years for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation
most importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC
the Court recognizes that Section 5 is inherently ambiguous. It is, therefore
up to the FTC to choose the legal standard under which
conduct will be reviewed
This is precisely the standard established by Chevron
The legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC
because the FTC has been reluctant to advance a more deferential standard
shell-shocked as it is from pre-Chevron losses
and has failed to recognize the current agency- deferential state of administrative law | Indiana
Dentists
is
cited for
de novo review
this is an over-reading
the case can
be
entirely in line with Chevron
Indiana
Dentists
had been argued before Chevron was decided
FTC was arguing for an uncontroversial interpretation
never
seek deference
had no need to argue for any
it would have been
surprising had
the Court cited
Chevron
today it is Chevron, not Indiana
Dentists, that is
the law of the land
for every regulatory agency other than
FTC
Section 5 is
ambiguous
therefore
the FTC
choose the legal standard
This is
the standard established by Chevron
the FTC has been
shell-shocked
from pre-Chevron losses
and
failed to recognize the current
deferential state of
law | 307 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: BUT, even before Chevron was decided, courts showed willingness to defer to FTC interpretations.\n\nHighlight Ratio: 0.4",
"role": "system"
},
{
"content": "Indiana Federation of Dentists bears special discussion. It is regularly cited for the proposition that courts conduct de novo review\nit seems remarkable that Indiana Federation of Dentists does not even mention Chevron—a fact that has led\ncommentators to believe that “[o]ne cannot explain judicial posture in the antitrust arena in Chevron terms\nthis is an over-reading\nIndeed, the case can instead be read as entirely in line with Chevron.\nit is unsurprising that Indiana Federation of Dentists does not cite Chevron. The Indiana Federation of Dentists petitioned for certiorari from a Seventh Circuit opinion that had been argued before Chevron was decided\nthe FTC was arguing for an uncontroversial interpretation of Section 5\nthe FTC had never structured its case to seek deference, and it had no need to argue for any deference\nGiven the case’s history and posture, it would have been more surprising had the parties or the Court cited to Chevron.\nMoreover, it took\nyears for the importance of Chevron to become understood and to filter its way into judicial review of agency statutory interpretation\nmost importantly, today it is Chevron, not Indiana Federation of Dentists, that is recognized as the law of the land—at least, for every regulatory agency other than the FTC\nthe Court recognizes that Section 5 is inherently ambiguous. It is, therefore\nup to the FTC to choose the legal standard under which\nconduct will be reviewed\nThis is precisely the standard established by Chevron\nThe legal issues presented are for the courts to resolve—but according to the legal standard prescribed by the FTC\nbecause the FTC has been reluctant to advance a more deferential standard\nshell-shocked as it is from pre-Chevron losses\nand has failed to recognize the current agency- deferential state of administrative law",
"role": "user"
},
{
"content": "Indiana\nDentists\nis\ncited for\nde novo review\nthis is an over-reading\nthe case can\nbe\nentirely in line with Chevron\nIndiana\nDentists\nhad been argued before Chevron was decided\nFTC was arguing for an uncontroversial interpretation\nnever\nseek deference\nhad no need to argue for any\nit would have been\nsurprising had\nthe Court cited\nChevron\ntoday it is Chevron, not Indiana\nDentists, that is\nthe law of the land\nfor every regulatory agency other than\nFTC\nSection 5 is\nambiguous\ntherefore\nthe FTC\nchoose the legal standard\nThis is\nthe standard established by Chevron\nthe FTC has been\nshell-shocked\nfrom pre-Chevron losses\nand\nfailed to recognize the current\ndeferential state of\nlaw",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: AND, recent clarification of FTC authority reassures the courts. | We generally agree with the Federal Trade Commission’s decision to rescind its 2015 Section 5 Policy Statement. Just as the Department of Justice and Federal Trade Commission Merger Guidelines are regularly updated on the basis of agency experience, legal and economic developments, so should this type of policy statement. Rescinding the old statement is particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.
Calls for reform have not come solely from Neo-Brandeisian commentators (including both FTC Chair, Lina Khan, and Tim Wu, now a member of the National Economic Council). The need for reform and a varied set of proposals has also been expressed by economics-oriented commentators, including this group of former Justice Department enforcers, Jonathan Baker and Herbert Hovenkamp, among others. Chair Khan in her statement suggested that the Commission would next consider replacing the Policy Statement with a new statement explaining how they plan to use Section 5 to increase competition. We think this would be a valuable way to show parties and courts what is coming. This article provides several suggestions that would be useful to consider and possibly include in the revised Section 5 Policy Statement. It should not be taken as an exhaustive list; there certainly may be other approaches to a revised statement that could also be effective.
A revised Policy Statement should make it clear that Section 5 is not identical to the Sherman and Clayton Act and that conduct can be challenged as an unfair method of competition under Section 5 even if it would not violate these other antitrust laws. In fact, even the original 2015 Policy Statement explicitly made this point. But the distinction between Section 5 and these other statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s cases and rules under Section 5 will likely face the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can help maximize the impact the FTC can have, while assuaging concerns of detractors who say there is no limiting principle. | the
decision to rescind its 2015 Section 5 Policy Statement
is particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.
replacing the
Statement with a new statement explaining how they plan to use Section 5 to increase competition
would be a valuable way to show parties and courts what is coming
the distinction between Section 5 and
other statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s
rules under Section 5 will
face the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can
maximize the impact the FTC can have
while assuaging concerns of detractors who say there is no limiting principle. | 308 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: AND, recent clarification of FTC authority reassures the courts.",
"role": "system"
},
{
"content": "We generally agree with the Federal Trade Commission’s decision to rescind its 2015 Section 5 Policy Statement. Just as the Department of Justice and Federal Trade Commission Merger Guidelines are regularly updated on the basis of agency experience, legal and economic developments, so should this type of policy statement. Rescinding the old statement is particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.\nCalls for reform have not come solely from Neo-Brandeisian commentators (including both FTC Chair, Lina Khan, and Tim Wu, now a member of the National Economic Council). The need for reform and a varied set of proposals has also been expressed by economics-oriented commentators, including this group of former Justice Department enforcers, Jonathan Baker and Herbert Hovenkamp, among others. Chair Khan in her statement suggested that the Commission would next consider replacing the Policy Statement with a new statement explaining how they plan to use Section 5 to increase competition. We think this would be a valuable way to show parties and courts what is coming. This article provides several suggestions that would be useful to consider and possibly include in the revised Section 5 Policy Statement. It should not be taken as an exhaustive list; there certainly may be other approaches to a revised statement that could also be effective.\nA revised Policy Statement should make it clear that Section 5 is not identical to the Sherman and Clayton Act and that conduct can be challenged as an unfair method of competition under Section 5 even if it would not violate these other antitrust laws. In fact, even the original 2015 Policy Statement explicitly made this point. But the distinction between Section 5 and these other statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s cases and rules under Section 5 will likely face the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can help maximize the impact the FTC can have, while assuaging concerns of detractors who say there is no limiting principle.",
"role": "user"
},
{
"content": "the\ndecision to rescind its 2015 Section 5 Policy Statement\nis particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.\nreplacing the\nStatement with a new statement explaining how they plan to use Section 5 to increase competition\nwould be a valuable way to show parties and courts what is coming\nthe distinction between Section 5 and\nother statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s\nrules under Section 5 will\nface the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can\nmaximize the impact the FTC can have\nwhile assuaging concerns of detractors who say there is no limiting principle.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: AND, recent clarification of FTC authority reassures the courts. | the
decision to rescind its 2015 Section 5 Policy Statement
is particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.
replacing the
Statement with a new statement explaining how they plan to use Section 5 to increase competition
would be a valuable way to show parties and courts what is coming
the distinction between Section 5 and
other statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s
rules under Section 5 will
face the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can
maximize the impact the FTC can have
while assuaging concerns of detractors who say there is no limiting principle. | 2015 Section 5 Policy Statement
particularly
hurdles
effective antitrust enforcement.
new
explaining how
use Section 5 to increase competition
valuable way
what is coming
other statutes
ignored
suppressed
rules
will
scrutiny of the courts
that
particularly helpful to have a clear Policy Statement
maximize the impact the FTC can have
assuaging
detractors
no limit | 309 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: AND, recent clarification of FTC authority reassures the courts.",
"role": "system"
},
{
"content": "the\ndecision to rescind its 2015 Section 5 Policy Statement\nis particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.\nreplacing the\nStatement with a new statement explaining how they plan to use Section 5 to increase competition\nwould be a valuable way to show parties and courts what is coming\nthe distinction between Section 5 and\nother statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s\nrules under Section 5 will\nface the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can\nmaximize the impact the FTC can have\nwhile assuaging concerns of detractors who say there is no limiting principle.",
"role": "user"
},
{
"content": "2015 Section 5 Policy Statement\nparticularly\nhurdles\neffective antitrust enforcement.\nnew\nexplaining how\nuse Section 5 to increase competition\nvaluable way\nwhat is coming\nother statutes\nignored\nsuppressed\nrules\nwill\nscrutiny of the courts\nthat\nparticularly helpful to have a clear Policy Statement\nmaximize the impact the FTC can have\nassuaging\ndetractors\nno limit",
"role": "assistant"
}
] |
Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio.
Tag: AND, recent clarification of FTC authority reassures the courts.
Highlight Ratio: 0.3 | the
decision to rescind its 2015 Section 5 Policy Statement
is particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.
replacing the
Statement with a new statement explaining how they plan to use Section 5 to increase competition
would be a valuable way to show parties and courts what is coming
the distinction between Section 5 and
other statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s
rules under Section 5 will
face the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can
maximize the impact the FTC can have
while assuaging concerns of detractors who say there is no limiting principle. | a new statement explaining how they plan to use Section 5
would be a valuable way to show
courts what is coming
rules under Section 5 will
face
scrutiny
At that time, it
help
to have a clear Policy Statement
This
maximize the impact
FTC
have
assuaging
detractors who say there is no limit | 310 | [
{
"content": "Based on the argument in the tag, highlight the text to support the argument. Highlight approximately the amount of text specified by the highlight ratio. \n\nTag: AND, recent clarification of FTC authority reassures the courts.\n\nHighlight Ratio: 0.3",
"role": "system"
},
{
"content": "the\ndecision to rescind its 2015 Section 5 Policy Statement\nis particularly relevant in light of the growing recognition of the hurdles preventing effective antitrust enforcement.\nreplacing the\nStatement with a new statement explaining how they plan to use Section 5 to increase competition\nwould be a valuable way to show parties and courts what is coming\nthe distinction between Section 5 and\nother statutes is often ignored or suppressed by commentators who object to more vigorous antitrust enforcement by the FTC. Eventually, the FTC’s\nrules under Section 5 will\nface the scrutiny of the courts. At that time, it may be particularly helpful to have a clear Policy Statement of how the FTC is interpreting Section 5. This can\nmaximize the impact the FTC can have\nwhile assuaging concerns of detractors who say there is no limiting principle.",
"role": "user"
},
{
"content": "a new statement explaining how they plan to use Section 5\nwould be a valuable way to show\ncourts what is coming\nrules under Section 5 will\nface\nscrutiny\nAt that time, it\nhelp\nto have a clear Policy Statement\nThis\nmaximize the impact\nFTC\nhave\nassuaging\ndetractors who say there is no limit",
"role": "assistant"
}
] |
Based on the argument in the tag, underline the text to support the argument.
Tag: Section 5 rulemaking with Chevron deference can reach extraterritorially. | 1. Hart-Scott-Rodino
The FTC’s regulations implementing HSR307 provide a good example of an agency using general rulemaking authority to define the geographic scope of a statute. HSR prohibits mergers and acquisitions that exceed certain thresholds, unless notification is given to the Department of Justice and the FTC and a waiting period has expired.308 The aim is to give these agencies the chance to review large mergers and acquisitions before they occur in order to determine whether they would violate the Clayton Act309 or the FTC Act.310 [FOOTNOTE 310 BEGINS] 310. Id. § 45(a) (prohibiting “[u]nfair methods of competition” and “unfair or deceptive acts or practices in or affecting commerce”). If the Department of Justice or FTC concludes that an acquisition would violate the Clayton or FTC Acts, it may seek to enjoin the acquisition. Id. § 18a(f). [FOOTNOTE 310 ENDS] HSR’s text does not define its geographic scope, instead speaking broadly of persons “engaged in commerce or in any activity affecting commerce.”311 As the FTC acknowledged when it issued its first regulations under HSR, the Act “contains no special provision for transactions having foreign aspects.”312
HSR expressly gives the FTC, with the concurrence of the Justice Department, rulemaking authority to implement the Act.313 [FOOTNOTE 313 BEGINS] 313. 15 U.S.C. § 18a(d). Such delegations are rare in antitrust law, although one author has argued for greater deference to administrative agencies in the antitrust context. See generally Justin (Gus) Hurwitz, Administrative Antitrust, 21 GEO.MASON L.REV. 1191 (2014) (arguing that the Supreme Court is moving towards bringing antitrust law within normal administrative law jurisprudence). [FOOTNOTE 313 ENDS] This authority includes the power to “define the terms used in this section;”314 to “exempt, from the requirements of this section, classes of persons, acquisitions, transfers, or transactions which are not likely to violate the antitrust laws;”315 and to “prescribe such other rules as may be necessary and appropriate to carry out the purposes of this section.”316 Beyond the authority contained in these grants, there is no express delegation to the FTC of authority to define the geographic scope of HSR.
The FTC issued its first regulations implementing HSR in 1978.317 From the start it used its rulemaking authority to define the Act’s geographic scope.318 It defined “commerce” to have the same meaning as in the Clayton and FTC Acts.319 The FTC explained in the Federal Register that because HSR’s commerce criterion “requires only that either the acquiring or the acquired person be engaged in commerce or in any activity affecting commerce,” HSR “permits coverage of a great many transactions that have some or even predominant foreign aspects.”320 To avoid HSR’s application to foreign transactions “with only a minimal impact on United States commerce,”321 the FTC created three foreign commerce exemptions: one for the acquisition of foreign assets;322 a second for the acquisition of voting securities of a foreign issuer;323 and a third for acquisitions by or from foreign government entities.324
Although these regulations are phrased as exemptions, their practical effect is actually to subject many acquisitions of assets outside the United States and many acquisitions of securities in foreign companies to the requirements of HSR. The foreign assets exemption, for example, provides that HSR shall not apply to the acquisition of “assets located outside the United States . . . unless the foreign assets the acquiring person would hold as a result of the acquisition generated sales in or into the U.S. exceeding $50 million (as adjusted) during the acquired person’s most recent fiscal year.”325 In other words, the FTC’s regulation provides that HSR does apply to assets located in foreign countries if those assets generated U.S. sales above a certain threshold. This interpretation holds even when both the buyer and the seller are foreign.326 The foreign securities exemption similarly provides that HSR does apply to the acquisition of securities in foreign issuers whose U.S. assets or U.S. sales exceed certain thresholds whether the buyer is a U.S. person327 or a foreign person.328
It seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine the geographic scope of HSR for itself, whether by applying the presumption against extraterritoriality or otherwise.329 As a general matter, FTC regulations implementing HSR are evaluated for reasonableness at Chevron step two.330 The D.C. Circuit has noted that HSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws.”331
In determining the geographic scope of HSR, the FTC did not apply the presumption against extraterritoriality.332 But the FTC did consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions whose principal impact is foreign.”333 In other words, the FTC’s regulations implementing HSR reflect the same values—domestic conditions and comity—that lie behind the presumption against extraterritoriality. Under the fourth model for evaluating reasonableness discussed above,334 the fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two. | The FTC’s regulations
provide a good example of an agency using general rulemaking authority to define the geographic scope of a statute
The aim is to give
agencies the chance to review
in order to determine whether they would violate
Clayton
or the FTC A
prohibiting
u
m
c
text does not define
geographic scope
gives the FTC
rulemaking authority to implement the Act
Such delegations are rare in antitrust
although
greater deference to
agencies in
antitrust
the Supreme Court is moving towards
authority includes the power to “define
terms used in this section;”
Beyond
authority
in these grants, there is no express delegation to the FTC of authority to define
geographic scope
From the start it used its rulemaking authority to define the Act’s geographic scope
It seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine
geographic scope
for itself, whether by applying the presumption against extraterritoriality or otherwise
As a general matter, FTC regulations
are evaluated for reasonableness at Chevron step two
HSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws
In determining the geographic scope
the FTC did not apply the presumption against extraterritoriality
But
did consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions
the FTC’s regulations implementing HSR reflect the same values
that lie behind the presumption against extraterritoriality
the fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two. | 311 | [
{
"content": "Based on the argument in the tag, underline the text to support the argument. \n\nTag: Section 5 rulemaking with Chevron deference can reach extraterritorially.",
"role": "system"
},
{
"content": "1. Hart-Scott-Rodino\nThe FTC’s regulations implementing HSR307 provide a good example of an agency using general rulemaking authority to define the geographic scope of a statute. HSR prohibits mergers and acquisitions that exceed certain thresholds, unless notification is given to the Department of Justice and the FTC and a waiting period has expired.308 The aim is to give these agencies the chance to review large mergers and acquisitions before they occur in order to determine whether they would violate the Clayton Act309 or the FTC Act.310 [FOOTNOTE 310 BEGINS] 310. Id. § 45(a) (prohibiting “[u]nfair methods of competition” and “unfair or deceptive acts or practices in or affecting commerce”). If the Department of Justice or FTC concludes that an acquisition would violate the Clayton or FTC Acts, it may seek to enjoin the acquisition. Id. § 18a(f). [FOOTNOTE 310 ENDS] HSR’s text does not define its geographic scope, instead speaking broadly of persons “engaged in commerce or in any activity affecting commerce.”311 As the FTC acknowledged when it issued its first regulations under HSR, the Act “contains no special provision for transactions having foreign aspects.”312\nHSR expressly gives the FTC, with the concurrence of the Justice Department, rulemaking authority to implement the Act.313 [FOOTNOTE 313 BEGINS] 313. 15 U.S.C. § 18a(d). Such delegations are rare in antitrust law, although one author has argued for greater deference to administrative agencies in the antitrust context. See generally Justin (Gus) Hurwitz, Administrative Antitrust, 21 GEO.MASON L.REV. 1191 (2014) (arguing that the Supreme Court is moving towards bringing antitrust law within normal administrative law jurisprudence). [FOOTNOTE 313 ENDS] This authority includes the power to “define the terms used in this section;”314 to “exempt, from the requirements of this section, classes of persons, acquisitions, transfers, or transactions which are not likely to violate the antitrust laws;”315 and to “prescribe such other rules as may be necessary and appropriate to carry out the purposes of this section.”316 Beyond the authority contained in these grants, there is no express delegation to the FTC of authority to define the geographic scope of HSR. \nThe FTC issued its first regulations implementing HSR in 1978.317 From the start it used its rulemaking authority to define the Act’s geographic scope.318 It defined “commerce” to have the same meaning as in the Clayton and FTC Acts.319 The FTC explained in the Federal Register that because HSR’s commerce criterion “requires only that either the acquiring or the acquired person be engaged in commerce or in any activity affecting commerce,” HSR “permits coverage of a great many transactions that have some or even predominant foreign aspects.”320 To avoid HSR’s application to foreign transactions “with only a minimal impact on United States commerce,”321 the FTC created three foreign commerce exemptions: one for the acquisition of foreign assets;322 a second for the acquisition of voting securities of a foreign issuer;323 and a third for acquisitions by or from foreign government entities.324\nAlthough these regulations are phrased as exemptions, their practical effect is actually to subject many acquisitions of assets outside the United States and many acquisitions of securities in foreign companies to the requirements of HSR. The foreign assets exemption, for example, provides that HSR shall not apply to the acquisition of “assets located outside the United States . . . unless the foreign assets the acquiring person would hold as a result of the acquisition generated sales in or into the U.S. exceeding $50 million (as adjusted) during the acquired person’s most recent fiscal year.”325 In other words, the FTC’s regulation provides that HSR does apply to assets located in foreign countries if those assets generated U.S. sales above a certain threshold. This interpretation holds even when both the buyer and the seller are foreign.326 The foreign securities exemption similarly provides that HSR does apply to the acquisition of securities in foreign issuers whose U.S. assets or U.S. sales exceed certain thresholds whether the buyer is a U.S. person327 or a foreign person.328\nIt seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine the geographic scope of HSR for itself, whether by applying the presumption against extraterritoriality or otherwise.329 As a general matter, FTC regulations implementing HSR are evaluated for reasonableness at Chevron step two.330 The D.C. Circuit has noted that HSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws.”331\nIn determining the geographic scope of HSR, the FTC did not apply the presumption against extraterritoriality.332 But the FTC did consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions whose principal impact is foreign.”333 In other words, the FTC’s regulations implementing HSR reflect the same values—domestic conditions and comity—that lie behind the presumption against extraterritoriality. Under the fourth model for evaluating reasonableness discussed above,334 the fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two. ",
"role": "user"
},
{
"content": "The FTC’s regulations\nprovide a good example of an agency using general rulemaking authority to define the geographic scope of a statute\nThe aim is to give\nagencies the chance to review\nin order to determine whether they would violate\nClayton\nor the FTC A\nprohibiting\nu\nm\nc\ntext does not define\ngeographic scope\ngives the FTC\nrulemaking authority to implement the Act\nSuch delegations are rare in antitrust\nalthough\ngreater deference to\nagencies in\nantitrust\nthe Supreme Court is moving towards\nauthority includes the power to “define\nterms used in this section;”\nBeyond\nauthority\nin these grants, there is no express delegation to the FTC of authority to define\ngeographic scope\nFrom the start it used its rulemaking authority to define the Act’s geographic scope\nIt seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine\ngeographic scope\nfor itself, whether by applying the presumption against extraterritoriality or otherwise\nAs a general matter, FTC regulations\nare evaluated for reasonableness at Chevron step two\nHSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws\nIn determining the geographic scope\nthe FTC did not apply the presumption against extraterritoriality\nBut\ndid consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions\nthe FTC’s regulations implementing HSR reflect the same values\nthat lie behind the presumption against extraterritoriality\nthe fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two.",
"role": "assistant"
}
] |
Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument.
Tag: Section 5 rulemaking with Chevron deference can reach extraterritorially. | The FTC’s regulations
provide a good example of an agency using general rulemaking authority to define the geographic scope of a statute
The aim is to give
agencies the chance to review
in order to determine whether they would violate
Clayton
or the FTC A
prohibiting
u
m
c
text does not define
geographic scope
gives the FTC
rulemaking authority to implement the Act
Such delegations are rare in antitrust
although
greater deference to
agencies in
antitrust
the Supreme Court is moving towards
authority includes the power to “define
terms used in this section;”
Beyond
authority
in these grants, there is no express delegation to the FTC of authority to define
geographic scope
From the start it used its rulemaking authority to define the Act’s geographic scope
It seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine
geographic scope
for itself, whether by applying the presumption against extraterritoriality or otherwise
As a general matter, FTC regulations
are evaluated for reasonableness at Chevron step two
HSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws
In determining the geographic scope
the FTC did not apply the presumption against extraterritoriality
But
did consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions
the FTC’s regulations implementing HSR reflect the same values
that lie behind the presumption against extraterritoriality
the fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two. | FTC’s reg
s
general rulemaking authority to define the geographic scope of a statute
review
Clayton
FTC A
u
m
c
not
geographic scope
FTC
rulemaking
implement the Act
rare
great
deference
agencies
antitrust
moving towards
authority includes
define
terms used in this section
no express delegation
geographic scope
start
court
geographic scope of HSR
FTC’s guidance
landed in exactly the same place
also
doubtful
cast aside the FTC’s guidance and seek to determine
geographic scope
for itself
presumption
otherwise
reg
s
reasonableness at Chevron step two
great discretion
not
did
commerce
comity
same values
behind
should be sufficient to find its reg
s reasonable at Chevron step two. | 312 | [
{
"content": "Based on the argument in the tag, emphasize the rhetorically powerful parts of the text to support the argument. \n\nTag: Section 5 rulemaking with Chevron deference can reach extraterritorially.",
"role": "system"
},
{
"content": "The FTC’s regulations\nprovide a good example of an agency using general rulemaking authority to define the geographic scope of a statute\nThe aim is to give\nagencies the chance to review\nin order to determine whether they would violate\nClayton\nor the FTC A\nprohibiting\nu\nm\nc\ntext does not define\ngeographic scope\ngives the FTC\nrulemaking authority to implement the Act\nSuch delegations are rare in antitrust\nalthough\ngreater deference to\nagencies in\nantitrust\nthe Supreme Court is moving towards\nauthority includes the power to “define\nterms used in this section;”\nBeyond\nauthority\nin these grants, there is no express delegation to the FTC of authority to define\ngeographic scope\nFrom the start it used its rulemaking authority to define the Act’s geographic scope\nIt seems doubtful that a court asked to determine the geographic scope of HSR without the benefit of the FTC’s guidance would have landed in exactly the same place. But it also seems doubtful that a court would cast aside the FTC’s guidance and seek to determine\ngeographic scope\nfor itself, whether by applying the presumption against extraterritoriality or otherwise\nAs a general matter, FTC regulations\nare evaluated for reasonableness at Chevron step two\nHSR gives the FTC “great discretion to define statutory terms and to promulgate rules to facilitate Government identification of mergers and acquisitions likely to violate federal antitrust laws\nIn determining the geographic scope\nthe FTC did not apply the presumption against extraterritoriality\nBut\ndid consider both the “impact on United States commerce” and “considerations of comity” when it decided to exempt “some acquisitions\nthe FTC’s regulations implementing HSR reflect the same values\nthat lie behind the presumption against extraterritoriality\nthe fact that the agency took into account the normative values underlying the presumption should be sufficient to find its regulations reasonable at Chevron step two.",
"role": "user"
},
{
"content": "FTC’s reg\ns\ngeneral rulemaking authority to define the geographic scope of a statute\nreview\nClayton\nFTC A\nu\nm\nc\nnot\ngeographic scope\nFTC\nrulemaking\nimplement the Act\nrare\ngreat\ndeference\nagencies\nantitrust\nmoving towards\nauthority includes\ndefine\nterms used in this section\nno express delegation\ngeographic scope\nstart\ncourt\ngeographic scope of HSR\nFTC’s guidance\nlanded in exactly the same place\nalso\ndoubtful\ncast aside the FTC’s guidance and seek to determine\ngeographic scope\nfor itself\npresumption\notherwise\nreg\ns\nreasonableness at Chevron step two\ngreat discretion\nnot\ndid\ncommerce\ncomity\nsame values\nbehind\nshould be sufficient to find its reg\ns reasonable at Chevron step two.",
"role": "assistant"
}
] |