The article points out that the energy to train BERT is comparable to a flight across America. There are hundreds of thousands of flights every day - why should we be concerned about the equivalent of one extra? Especially given that BERT improves ~10% of English language Google search results, it seems like we're getting a lot in return for relatively small energy use. On top of that, Google buys 100% of their energy usage from Green sources.
I think it's great to talk about other methods of training or architectures that don't require so many parameters. The point about how BERT consumes vastly more text than humans do when they learn to read is interesting. But trying to phrase this like an environmental issue just seems disingenuous and misleading.
Not sure why folks are defensive? The article is mainly informational.
TFA:
"What does this mean for the future of AI research? Things may not be as bleak as they look. The cost of training might come down as more efficient training methods are invented. Similarly, while data center energy use was predicted to explode in recent years, this has not happened due to improvements in data center efficiency, more efficient hardware and cooling."
Also. One flight in and of itself isn't an environemntal problem; thousands of flights are.
Training this one instance of a model isn't an environmental problem; I would be curious to see some educated guesses about the number of models being trained over the next ten years. Not an expert but I know that use of ML is exploding - lots of new use-cases and thus lots of new models.
The article suggests that the energy usage of language models is a problem. I don't think energy usage is a problem. I'm not sure how you interpret this as being defensive.
There are hundreds of thousands of flights per day. Adding an additional flight, or even an additional thousand flights, to substantially improve Google doesn't seem like a big cost. Consider: Would your life be worse if one random trans-American flight was cancelled today, or if Google searches became 10% worse?
Another way of thinking about this same point, is, why is the author writing about language models if they are so concerned about the environment. Surely the airline industry is a better subject as, again, they fly hundreds of thousands of flights each day. It's hard to take someone seriously when they are focusing on an infinitesimal part of a huge problem.
It's also misleading because there is a difference between energy used by an airline flight, where the energy comes from burning jet fuel, and energy used in a data center. In Google's case (the people training BERT) the energy used was 100% renewable - Google reached that goal in 2017. Perhaps Open AI didn't use renewable energy to train GPT-3, but I wager they didn't power their machines by burning jet fuel either.
Maybe electricity used to power to train language models will become a meaningful issue at some point in the future. I don't think that future is close at hand though.
tl;dr I think the tech industry doesn't like to be criticized.
That's a good point about renewable energy sources.
I'd still argue it's reasonable to assume that ML will consume more power as use-cases for it grow (i.e. more and more models are trained), and, therefore, I don't think it's unreasonable to consider the energy usage.
Honestly - I think what you may have heard is "ML is bad/evil." And yeah some folks probably feel/vibe that.
That's different from "Like any other technology the full context of ML deserves to be considered, and since it's so hyped by billion-dollar companies right now, maybe a bit of pushback on all the hype isn't a terrible idea."
And yeah we should definitely look at the airline industry; pretty sure Mother Jones does, and there's no reason we can't look at the energy impact of multiple industries.
If you were against lung cancer, and writing about the issues caused by scented candles emitting smoke and that smoke corrupting lungs - it would be hard to take you seriously or accept your criticism as valid. It's probably true that scented candles are infinitesimal contributors to lung cancer, but it's just silly to think of them that way.
Likewise, yes, language models use energy, but objectively, it's not that much energy and the sources for the energy are sometimes green and always efficient (i.e. Coming from the grid and not by burning jet fuel).
You can't really solve the problem of lung cancer by addressing scented candles. Reducing energy used by language models likewise won't have an effect.
As always, there’s nuance. You can schedule workloads, machine learning/training included, where the power is cleanest (low carbon). Google relies on electricitymap.org/Tomorrow to do this.
The more folks who elect to pick where to compute based on low electrical carbon intensity, the faster the grid turns over to clean generation. You must vote with your fiat. I encourage technologists to include this consideration in their workload scheduling requirements. Renewables are almost always cheaper than fossil generation as well.
The categorical imperative doesn't require everyone to be practically able to do the thing. Is it not moral to feed a starving man because people in China aren't able to feed him?
There's more than enough solar and wind, it's just a matter of how much we as a society want it. The green energy revolution has barely begun, and we're still climbing the learning curve. More demand means a faster climb means exponentially more green energy sooner.
We’re already at soon. Another order of magnitude increase (which at our pace is all but guaranteed in the near term) will make large scale training a significant R&D line item even for the FAANGs. It will also push efforts to be commercially viable, as you can longer do proof-of-concept nets at $500m
> I think the point is, if nothing changes, the costs will be very significant very soon.
The energy efficiency of machine learning hardware is progressing at a rapid rate. It's not accurate to assume that the energy costs will stay the same. Just look at how much it would have taken to train something like GPT-3 five years ago.
I think it is reasonable to assume that any advances in efficiency will be subsumed by increased use, much in the same way increased energy efficiency doesn't decrease absolute overall use.
ML hardware is no doubt progressing. Whether we can expect a rapid rate is a different matter. The various Moore's Laws have been breaking down (chip speed stopped doubling a while back). GPU are the sort of chip that could most benefit from just transistors doubling (last Mooore's law standing) but even this Moore's Law is breaking down.
ML model size has had it's own Moore's Law, with standard model growing exponentially in size [1]. And this implies model are going butt up against the limits even more than they have already. Whether the "bitter lesson"[2] of ML is true inherently is up for a question. That current researchers have accepted it seems given.
My brain does nlp better than any system out there. I’m also able to ride bicycles and do motor control better than Boston Dynamics. I can also construct and prove math, do physics and code all of this in Matlab and C. My brain can handle all these wide range of tasks almost seamlessly with just 15 watts of power, that Silicon Valley’s super computers can barely do 1% of.
To be fair, it takes years of energy to train our brains to the point where it can do all those things, and our brains aren't extensible in the same way hardware/software is. I guess there's also a lot more variation in yields. ;)
Well, that’s not entirely true. Alpha go self played 29 million games of Go, which is in feasible for humans. Assuming a game of Go is 5mins, it would take a human 286 years to achieve the same, assuming this person doesn’t sleep, eat or do anything else other than just play Go. CPU time is an order of magnitude faster than real time, especially on GPU clusters.
To be a fair comparison wouldn’t you have to include all the energy used to train “your” brain through generations of evolutionary training? Your latest model is like taking an already trained BERT model and adding a few tweaks
I see BERT as non sensical. You need to be scientific and have a mathematical theory on how humans learn language, which is a multi disciplinary task requiring physicists, mathematicians, neuroscientists, cognitive psychologists and linguists. Benchmarks are useless, theories, models, experiments and testable predictions is how science progresses. You’re making a comment on cognitive science, and trying to imply that language learning in humans isn’t learned, but pre baked. The psychological, linguistic, evolutionary biology and neuroscience evidence doesn’t seem to corroborate. The evidence points stronger to humans having general learning and problem solving abilities. For instance, there was no evolutionary pressure for humans to be good at math or programming. I was not born knowing english or calculus or probability theory, these were learned abilities. Evolution favoured brain mechanisms that lead to behaviour for success in a rapidly changing world. Had I been born in ancient Rome as a farmer, I would learn to speak Latin, and learn how to be a successful farmer, instead of the physics, math, probability, computer, driving, reading skills that I learned in my life time.
>You’re making a comment on cognitive science, and trying to imply that language learning in humans isn’t learned, but pre baked.
You make good points, but this one isn’t quite what I meant. I didn’t mean we are trained by evolution for a particular language, but that evolution selected for a language skill. In other words, we are “pre-baked” with the ability to learn a language. It would be analogous to a DL model being trained for generalized regression but not a particular problem. To that extent, I think we’re saying the same thing. Some of the theories related to our generalized learning abilities postulate they stem from this base ability (our aptitude for music, for example, being a consequence of our language learning ability)
>I see BERT as non sensical. You need to be scientific and have a mathematical theory
This is a matter of contention. A lot of science progresses by starting with an empirical result that drives a change in theory rather than the other way around. I can’t remember who to attribute it to, but there’s a quote to the effect that “‘Thats odd’ is the most productive phrase in science, rather than ‘Eureka!’”
Yes it did, and we do know a lot about general principles behind, from different disciplines. This is still an early science. NLP research still hasn’t considered many important aspects of language learning that we have discovered in such a short period of time.
> starting with an empirical result
What makes you think these benchmarks are empirical? They were hand constructed to fit some objective, assuming that being good at the said objective is required for NLP tasks. Where’s the empirical experiments to validate the notion that said objectives lead to language? Science hasn’t worked this way, datasets aren’t constructed, you do an experiment and MEASURE it. Then you make models, try to explain the phenomenon, and test new ideas and validate your models. Can your model extrapolate new information and suggest new experiments to validate? I used the word extrapolate over predict intentionally.
It’s a little hard for me to follow your last paragraph but it sounds like you are confusing AI and AGI. I don’t think anybody using BERT is claiming the latter.
I assume the empirical results are the validation sets run. I.e., the tests that show it provides better results than the base rate. Again, it’s important not to conflate verifiable results with understanding the underpinnings of why it works. If you’re walking and I race you with my car over and over again, I can conclude my car is a faster mode off transportation without understanding anything other than “push right peddle to go faster”. My ignorance doesn’t invalidate the results
Yeah but your clone() function is very wasteful, and we can't stick 1 million of you in a dark room that just looks at people's google searches and figures out what they're really going for.
And it's not the whole picture to say the brain only uses 15 watts - when there's all sorts of necessary support systems that it couldn't run without. So it's closer to 100 watts (2000kcal/day)
Not Google search, but many companies are exploiting cheap labor over seas for different industrial use cases because our algos suck. Not just in manufacturing, there are tech companies outsourcing labeled data for tasks like object detection (Hive.ai as anvexample). Whether it makes you depressed or not, humans are cheaper than algorithms, and I don’t see that changing unless we abandon the current paradigm for ai/ml.
Comparing your pre-trained brain (that has a lot of structure (=training) through evolution) with the training costs of a new algorithm isn't really fair.
All in you require about 100W (2000kcal/day), maybe 3 times that when doing a lot of physical activity. Boston Dynamic's Spot uses about 400W. I can probably outperform it in some disciplines while it would beat me in some others. That would be a fair comparison.
You’re making sweeping assertions that require domain experts in several different disciplines. The scientific evidence points to the contrary. It is demonstrated that humans have a general ability to learn wide range of things, without being genetically programmed to. None of us evolved to drive cars. Yet nearly everyone in my grandparents generation were able to learn this totally new skill despite being a new invention, where you couldn’t possibly have had time for evolution to act. They weren’t genetically evolved to drive, it was learned within their lifetime and generation. There are tribal humans in different parts of the world that haven’t developed written language. Yet you can teach them written language. Where’s your “pre trained brain” theory there?
Your ancestors might not have driven cars, but they used the motoric skills required for controlling one, and they had a vision system optimized for detecting and avoiding moving objects.
Human brains are also the culmination of billions of years of iterative development. Computers were only invented in the last century. I would not be surprised if computers could catch up given a few tens or hundred more years.
> I'd give 30(realistically 15) years tops for AGI.
I think 30 years is probably reasonable, if for “30 years” one reads “twice as long from now as commercially viable fusion generation actually was when first widely hailed as 15 years away”.
I think this is false. My general opinion of computer scientists and engineers (especially in silicon valley) lack the scientific training that one gets in other disciplines like in physics or other hard sciences. Psychology, linguistics, cognitive science and neuroscience has generated a rich diverse experimental data in the past 50 years, and now is the prime time for a theory to emerge to connect everything. To make what I’m saying clearer, you needed Newtonian mechanics, Maxwell and Plank and Coulomb to develop and discover the science of electricity and atoms in order for us to engineer the silicon transistor. Without the original scientific knowledge, building such devices would have been impossible. This machine learning, ai, deep learning and this obsession with benchmarks are, in my opinion, a hinderance to AGI
I think the error in your thinking is the assumption that scientific discovery is predicated on previously defined theory.
Theory is required for understanding, not for discovery. They are two different things. What gets a lot of scientists and statisticians worked up is how ML can outperform traditional models without always being interpretable. It’s like claiming the Wright brothers couldn’t build a flying machine without having a thorough understanding of the mechanics of flight. I can improve performance of my car by remapping the fuel without necessarily understanding the nuances of optimizing enthalpy. There’s levels of understanding, and they may not correlate completely to effectiveness. To that extent, it’s more engineering than science
> There are hundreds of thousands of flights every day - why should we be concerned about the equivalent of one extra
See also: Sorites fallacy.
There is a limited desire for flights, which benefit people as far as it allows them to get things where they need to be, and no more than that; there is presumably an unlimited desire for machine learning.
> One recent model called Bidirectional Encoder Representations from Transformers (BERT) used 3.3 billion words from English books and Wikipedia articles. Moreover, during training BERT read this data set not once, but 40 times. To compare, an average child learning to talk might hear 45 million words by age five, 3,000 times fewer than BERT.
The human brain is the output of an incredible number of generations of training, representing a vast consumption of energy. Most of the learning that informs the brain happened before this hypothetical five-year-old was even born.
That is a matter of great debate. Although most people's brains wind up settling on basically identical structure over time, case studies with injured or disabled people often display incredible adaptability (ie the occipital lobe in a blind person). If indeed the brain learns most of what it knows from experience, than its (comparatively) low energy consumption would come from greater efficiency. That definitely seems possible; as a basic unit of machine learning the neuron is much more specialized than the transistor, and far slower.
My interpretation of what the parent meant is that the cognitive processes which support language acquisition (which may be mostly domain-general) are already optimized for this task, for which there is a sensitive period during certain years where you need to be exposed to certain inputs, or else you never fully acquire language. See: https://en.wikipedia.org/wiki/Genie_(feral_child)
550–600 million years of hyperparameter tuning and neural architecture search using an evolutionary algorithm is nothing to laugh at.
But the actual learning process of a single brain uses very little training examples and is very energy efficient. A brain that uses only 20 watts of power.
This is beyond ridiculous, even if compared to the energy to train a human mind at the age of, let's say, 35. Raise a child to the age of 35 and the spent energy will surpass the mentioned energy cost by at least 2 orders of magnitude.
And no, I don't ridicule AI ethics researchers: the energy cost of AI is negligible to the serious issues. The real harm stems from other areas.
The 6000kWh number is assuming 20W over 35 years. That might be right if you only count the brain, but then you would have to revise the neural network numbers to only count CPU and memory, which would easily halve those numbers (fans take a lot of power).
If we want to account for power supply, temperature control etc. we could use the entire calorie count of a human over 35 years, leaving us with around 30000kWh (assuming a health, not-overweight human). You could now argue that that's too high (the human is doing a lot of other things). But as you pointed out it's a bit of a pointless comparison anyways as humans don't start from scratch.
> we could use the entire calorie count of a human over 35 years, leaving us with around 30000kWh (assuming a health, not-overweight human)
That's still too low, considering you can't just have a human eating food and educating themselves in a vacuum.
We need to account for energy going to and from school, shelter, heating, clothing. Basically all the necessities that enable someone to learn (and their professors)
There are very few "typical human tasks" that machine learning actually does better than a person at; playing game all the occurs to me. But driving, image description, language-translation, and so-forth, at the scale that a human does it, the human is generally still considered to do tremendously better. Of course, some ML programs are convenient in being able to do things at a large scale but a majority of what ML programs do is like language translation - not great and what we sometimes live with 'cause it's free.
Basically, one should really not discount human ability, especially in tasks considered mundane. Such skills often are often considered "mediocre" not because they're actually easy but because all human can do them, see:
ML is usually better at speed. For example I use a fine tuned GPT-2 for code autocomplete in vim, and even though humans could do better, they can't do it as fast.
Right now, a lot of the low-level tasks that contract lawyers do at scale for maybe $50/hour could be done by clerks for $15/hour or clerks in India for $3/hour (or something). But the lawyer paid for their time because their expertise is need to flag the "corner cases".
Fine tuning BERT (what most people in practice do) is so much more efficient. You can do it in 8 hours on a 2080ti.
They need to mention that the total number of people training large transformers from scratch is very very small. If wager that the total number of different, uniquely trained (not using previous weights - which reduces compute necessity by massive amounts) language models in existence is in the low hundreds
I'd claim that these mass language models serving as the underlying encoding backbone behind more specific systems actually save energy and compute compared to the previous methods (needing far more data and thus more energy spent on getting it combined with less efficient representations like tf-idf causing many classifers to perform very slowly and thus burn lots of energy)
Also, much of the recent research in this field is about model pruning, quantization, and any technique you can imagine to reduce training and inference time or memory requirements.
All in all, big language models are a net positive for the environment. The effeciency gains in any number of fields from increasingly sophisticated NLP systems far, far outweighs the costs or of training them. Foundational research in environmental conservation will be accelerated by effective NLP semantic search and question answering systems. That's a single, tiny example of the potential for benefits from large language models.
That's my #1 reason to be bearish on B$ - it's just a complete environmental cluster-fuck. And in order to not see this, you'll have to ignore a lot of facts - which in turn tells me a lot about those inside the crypto-bubble, namely that they do not care that much about facts.
at Starship's $100/kg placing your Eth mining $1000 GPU in space will add only 10-20% when amortized over 100 ton rig - it is comparable with building/buying your own small powerplant what large crypto have done. AI and crypto are exploding and only going more so. Granted, the planet is becoming too small a confines for it. It seems that AI and crypto will be the killer apps of space in near future.
One can also observe that humans have largest, among the animals of planet Earth, share of body energy consumed by brain and the future humans would probably have even higher share. In technology we observe the same - the pinnacle of technology - CPU - have practically 0 thermodynamic efficiency and the share of energy consumed by computers grows, and i think it will be only growing. Intelligence eats the world.
Heatsinks optimized for radiative cooling, which are kept in the shade by solar panels. Works much better when half or more of your surroundings aren't a planet at human-scale temperatures, as is the case on earth's surface.
only because of very constrained weight budget (which is constrained by the high price of delivery into space) so you can't just throw in an AC with a radiator. The radiative power scales with the 4th of T, so it is pretty effective in space if you run the hot end of AC hot.
I suppose there were a lot of arguments about usability and various constrains/impossibilities 30+ years ago when cell phones only appeared at $1+/minute and were available only at a very few places. Yet here we are. The cheap access to space will do the same for various tech-in-space (how about unlimited access to space with fineprinted "cap of 1000kg/month" :).
Could you clarify why you think placing crypto rigs in space is a good idea? To run a crypto rig, you need to provide a lot of power and have a mechanism to keep the processors cool. Being in space seems to make both of these substantially more difficult (especially cooling the electronics). It seems highly implausible to me that crypto will be the “killer app” of space.
My #1 reason is that once someone figures out how to build large, reliable quantum computers and program them in a practical way and quickly break sha256, bitcoin might be how we find out. That still seems to be a ways away though, and in the meantime, having a mostly-unbreakable digital ledger seems to be valuable.
The counterpoint to this I'll make is that all financial systems engage in costly signaling.
If you visit a foreign country how do you know to trust whether to trust a specific bank? You probably look at cues like whether it occupies an expensive skyscraper in the center of town? Do you see its ads around town? Does it sponsor the local soccer team? All credible signals that are hard to fake for a fly-by-night scammer.
All Bitcoin did was formalize this process. At any given time there are many chains that all purport to be the canonical history. How do you decide which one is authentic? By looking at hard-to-fake signals. In this case the accumulated hashing power behind the chain. Looking for whoever spent the most hash work is fundamentally no different than checking to see which bankers are wearing the most expensive suits.
Any system with trusted intermediaries will waste resources on costly signaling. The only question is whether crypto mining is more of a fundamental waste than traditional signals, like high-paid bankers and prestige real estate.
I don't know. I see the nice HSBC building in my city and all I can think of is that half of the business of theirs is crooked, money laundry, trafficking, drug cartels. No shiny building can change what you actually do.
> The counterpoint to this I'll make is that all financial systems engage in costly signaling.
Bitcoin uses 600+ kWh per transaction currently. With current use, to process VISA's 1700 transactions per second would consume >36 PWh annually. That's significantly higher than global electricity consumption and capacity. At current US electricity prices it would be significantly over 4.5 trillion USD annually. Global annual banking revenue is under 6 trillion: https://www.mckinsey.com/industries/financial-services/our-i...
> The only question is whether crypto mining is more of a fundamental waste than traditional signals, like high-paid bankers and prestige real estate.
You are assuming that power used by the network is directly proportional to the rate of transactions processed. I doubt that is a sound assumption.
For example miners use energy to compete for the mining reward which is awarded every ten minutes. I can think of many ways of modifying the network to increase the rate at which transactions can be processed that do not increase this mining reward. (I do not understand why none of those ways have been adopted.)
One of the primary reasons that reward is not lower is probably because the amount was set when Bitcoin was created, and people worry that changing it would set a precedent making other changes easier.
The amount of work is what keeps transactions secure. The difficulty of computation is directly proportional, but the power usage depends on the demand for mining.
Lighting attempts to decrease this load by effectively batching transactions, but its not a huge difference compared to the orders of magnitude differences that exist already.
the vast majority of bitcoin/crypto mining uses renewable or otherwise wasted energy and this is the only economical way to do it! you have been intentionally misinformed by the amount of energy used and not the source of energy used. It is completely a red herring to just read a headline about how little electricity a tiny country uses and that mining uses the same amount.
what? "wasted energy"? yeah a lot of energy is lost as it cannot be transported to commercial and residential areas economically. so miners set up processing at the source of the energy and use it. in fact, a lot of it actually reduces pollution creating the polar opposite of what you believe, in those circumstances it is a sustainability solution.
now aside from fighting me about your worldview, this is also an area to remain vigilant about! nation states can absolutely mine at a loss when they turn to competing for control of the cryptocurrency networks for geopolitical reasons. right now it is renewable.
but what about the hardware, the single purpose hardware and e-waste? there is a large trade of "obsolete" hardware, as it is economical to use for people with the cheapest power. this is also another specific area to be vigilant about, by making sure the infrastructure is in place to put that hardware and keep it active, instead of landfills.
I can see that mining is a good incentive to look for excess energy in all kinds of places and make use of it. That's a capitalistic motiv, finding solutions to derived issues, when something else is at the core of the problem. But anyway.
B$ still looks ridiculous, let me explain.
* world electrical energy consumption: ~25000 TWh [1]
* bitcoin energy consumption index: ~77 TWh [2]
Bitcoin today consumes as much as 0.3% of the total energy available on this planet.
That's fine, if half the world would use it and would do meaningful things with it.
What is it actually used for most visibly?
As a betting ground with galactic momentum, a technology promising to be the solution to everything, and just outrageous claims that only the bovine left to be excited about.
It's not that the algorithms and data structures are not cool, the certainly are - but we can do so much more today with technology than this.
And that's a different issue, you are conflating your energy use criticism with an arbitrary meaningful use criticism. Others may be willing to have that argument with you, but I know how those go: Someone says something about another industry's use, you say something about slow databases, they say something about another use case, you say something about why bitcoin/blockchains are not the best solution for that use case, rinse repeat.
The amount of energy without detailing the source of energy is not a valid argument to support any argument you have about why it is being used. Move off of it, or criticize the sources in a more nuanced educated argument. It should be 100% renewable or part of a sustainable solution, not only will you get further in your argument by focusing on the market participants that actually are wasting industry, you will also be helping!
The fact that it is primarily renewable is actually the reason why BTC's energy costs are exorbitant. That energy could be directed toward replacing fossil fuels, thus reducing carbon emissions.
You see, we have these things called power lines, and, along with the substations, transformers, and other transmission equipment, those make up power grids. We have 3 of them in the US alone.
Now, when electricity is generated, you can transmit it over those power grids, even to places that aren't right next to the dam. You can even store a limited amount for later use.
Ever kWh that gets redirected away from frivolities like BTC can go toward something that would otherwise be powered by fossil fuels.
I thought you were "educated" here? Please do keep up.
Au contraire, I appreciate your "educated" opinion here. I just want to make sure I'm sufficiently "educated," so I can believe the exact same thing you do, with a straight face. If you didn't want it that way, you shouldn't have led with it.
Regarding A), there's no particular reason why that electricity can't be sent across power lines.
B) So what? Send it anyway, and replace 85-92% of the energy use.
So you think that I'm not open to conflicting realities, got it. I am not speaking in absolutes I am simply responding to you. Now you know. I led with an "educational moment" because it still remains accurate that the amount of energy is not a strong argument without referencing the source of the energy. It's not controversial for you to disagree with that, a reality still was and now is with mining included that energy producers opted to not transport some energy across long distances and miners are overwhelmingly using that because it is economical for both parties. You read something else that I didn't say or imply. You keep referring to that, which I was willing to ignore. It's clear snark is important to you, now you got your reaction so I hope you feel satisfactorily accomplished in that regard.
Now let's talk about your unique points than the original person I replied to: places have had decades to send more power over power lines and they didn't. You tell me why that is. I assume there were financial reasons and related impracticalities. But admittedly I have never asked and only react to the reality that energy producers have been receptive to the additional use of their energy for the aforementioned reasons I listed. Their output hasn't changed and to them it is a more efficient use of it. Are they lying? Are they ignorant of alternatives? Just lazy even though their laziness would therefore predate crypto mining?
Either way mining on premise is an immediately applicable economic incentive that simply got you to notice that you didn't like it.
Best case scenario then is that it gets you into action to implement a solution nobody else noticed was applicable. Society might be getting somewhere because of you, that's so exciting.
Yes, very "exciting." I thought we were done with the snark?
Nonetheless, financial difficulties are not relevant. What is relevant is reducing the world's carbon footprint, and that needs to happen ASAP. I, personally, think it's more "exciting" that human civilization might survive another century than to keep track of the solution to some useless, financialized math problem. But, that's just me, I suppose.
Crypto mining at fracking sites does this. The flare gas systems throw hydrocarbons into the atmosphere before the miners arrived, the miners doing a financialized math problem give them the missing incentive to use the energy for the mining computers on site.
Corroborate that with a news source that you happen to like.
Like I mentioned renewable or otherwise wasted energy. This is one of the otherwise wasted energy examples that is a sustainability solution right now, in comparison to wishful thinking.
Shouldn't that be a reason to be bullish? Because if the governments step in and enact regulation against Bitcoin mining, that will increase its scarcity and therefore value.
Likewise, if you think fossil fuels will be scarce in the future, you can be bullish on the price of gas, but not bullish on things based on gas.
The number of bitcoins available (disregarding lost/forgotten ones) is fixed at 21M, and the rate at which they are created is more or less constant (by design), so anti-mining regulations will not increase scarcity. If anything, restricting the network hash rate will make the network less secure and thus less valuable
Not that I'm aware of. The lightning network seems to be a fairly decentral affair to me. Still I suggest you make up your own mind by perusing https://lightning.network/
So the arbitration still occurs on the blockchain? Seems like maintaining all these different ledgers adds a lot of complexity when a better solution may be a non-Bitcoin, or even non-blockchain DB.
It’s still crazy to think about how much more efficient computational power is nowadays compared to even five years ago. I remembered I changed my power supply in anticipation for Nvidias then newest GTX series. But only to learn that a 1050Ti only had a 75W max draw so the new fancy 1kW PSU was completely unnecessary. Also to put into perspective, my Mac mini has a 30W max draw. 30W!!! It’s ridiculous how little energy it draws.
The brain is not just more efficient, it also builds itself. The hardware necessary to run a neural net is given. Maybe the rigors of self replication are what's missing from AI.
Problem is Gebru is an activist, she's biased towards her own cause. Accepting her demands directly, without including all the other groups and voices, would not be right.
> This month, Google forced out a prominent AI ethics researcher after she voiced frustration with the company for making her withdraw a research paper. The paper pointed out the risks of language-processing artificial intelligence, the type used in Google Search and other text analysis products.
And they aren't even learning in a sense that would make them categorically comparable to human-like learning. You could just say "it takes a lot of energy for a computer to crunch a lot of numbers".
Why does the computer need to learn when we can just program what the computer is supposed to do based on what we know what the desired output/result should be? Yeah, we can try to recreate calculus from first principles, or just use the equations that we already know. unless I am missing something obvious. There are PHP or C scripts for example can do image recpnition and break captchas (there is a program called Xrummer that solves captchas for spamming purposes, which is why captchas have become so complicated) ...this was in 2010 , long before machine learning became a 'thing'.
As complicated as it is, deep learning is simple compared to programming by hand all the required knowledge. This approach historically failed (expert systems).
I don't know that a post this arrogant warrants a serious response. If you think you can write a PHP script that does what BERT or ImageNet does, by all means go right ahead.
I think most of the critics here has to come from people not understanding that GPT-3 was not and will not be trained more than a few times.
If even then you still want to argue, aluminum consumes 17,000kWh per ton ton which is 11 BERTs. I don't see weird criticism about mining consuming energy. Doing stuff takes energy and while the sentiment against ML seems to be "it's snake oil", it is still very real research with actual real-world impacts.
Finally, we are in a transition phase where ML is done on GPUs which consume more energy than dedicated ASICs. We already have Google TPUs and Amazon Inferentia which can be used today. Power consumption will go down as dedicated hardware gets better and better.
I think it's great to talk about other methods of training or architectures that don't require so many parameters. The point about how BERT consumes vastly more text than humans do when they learn to read is interesting. But trying to phrase this like an environmental issue just seems disingenuous and misleading.
reply