Why the Future of AI & Computers Will Be Analog
Summary
TLDRThe video script discusses the resurgence and potential of analog computing in a world dominated by digital technology. It highlights the energy efficiency of analog systems, which can be 1,000 times more efficient than digital counterparts, and how this could be part of the solution to the climate crisis. The script also touches on the limitations of digital computing, such as physical boundaries and energy consumption, and introduces companies like Mythic and Aspinity that are developing analog chips for modern applications. The potential for hybrid computers that combine the best of both worlds is also explored, hinting at a future where analog computing could play a significant role in our daily lives.
Takeaways
- đș Analog computing, once overshadowed by digital, is experiencing a resurgence due to its potential energy efficiency and unique problem-solving capabilities.
- đĄïž Analog systems have an infinite number of states compared to digital systems, which rely on a finite number of states determined by bits or transistors.
- đ The Space Age and personal computers marked a decline in the size of computing devices, but analog computing might be reaching physical limits in terms of miniaturization.
- đĄ Digital computing, particularly in areas like AI and cryptocurrencies, is increasingly energy-intensive, prompting interest in more efficient alternatives like analog computing.
- đ A return to analog computing could significantly reduce energy consumption, with analog processes sometimes being 1,000 times more efficient than digital ones.
- đ ïž Analog computers operate based on physical models that correspond to the values of the problem being solved, as opposed to digital computers that follow algorithms and discrete data.
- đ The limitations of digital computing are being recognized, with experts like Bernd Ulmann suggesting that we are approaching the fundamental physical boundaries of digital elements.
- đ§ Analog computing's continuous data processing allows for real-time problem-solving and efficient parallel processing without the need for cooling facilities.
- đ Hybrid computers that combine the energy efficiency of analog with the precision of digital are being explored for future technology development.
- đ Everyday applications of analog computing could include low-power sensors for voice-enabled devices, environmental monitoring, and wearable technology.
Q & A
What is the fundamental difference between analog and digital computing?
-Analog systems have an infinite number of states and can represent a continuous range of values, while digital systems rely on a finite number of states determined by the number of bits or transistors that can be switched on or off.
How has the advancement of digital computing impacted the size of computing devices?
-Digital computing has led to a significant reduction in the size of computing devices, from large machines to personal computers and smartphones, following the predictions of Moore's Law which suggests a doubling of transistors on integrated circuits approximately every two years.
What are some of the environmental concerns associated with digital computing?
-Digital computing, especially in data centers and power-hungry applications like cryptocurrencies and AI, is becoming increasingly energy-intensive, contributing to global energy consumption and carbon emissions. It also requires substantial cooling systems, which can strain water resources.
Why is analog computing considered more energy-efficient than digital computing?
-Analog computing can perform the same tasks as digital computing with a fraction of the energy because it operates on a physical model corresponding to the problem being solved, which doesn't require the switching of transistors and can handle continuous data in real time.
What is the significance of the MONIAC computer in the history of analog computing?
-The MONIAC (Monetary National Income Analogue Computer), created by economist Bill Phillips in 1949, is a classic example of analog computing. It was designed to simulate the Great British economy on a macro level using water to represent money flow, and it could function with an approximate accuracy of ±2%.
What are some practical applications of analog computing today?
-Practical applications of analog computing today include flight computers used by pilots for manual calculations, as well as emerging technologies like low-power sensors for voice-enabled wearables, sound detection systems, and heart rate monitors.
How does the concept of Amdahl's law relate to the limitations of digital computing?
-Amdahl's law suggests that the speedup of a system is limited by its sequential operations that cannot be parallelized. As a result, adding more processors does not always lead to proportional improvements in speed, which is a challenge for digital computers when trying to handle increasingly complex tasks efficiently.
What are some of the challenges in integrating analog and digital systems?
-Integrating analog and digital systems requires seamless connectivity and synchronization between the two paradigms, which can be technically challenging. It also involves developing hybrid computers that combine the energy efficiency of analog with the precision and flexibility of digital computing.
What is the potential impact of analog computing on machine learning and AI?
-Analog computing has the potential to significantly reduce the power consumption of machine learning and AI applications by offering a more energy-efficient computing method. Companies like Mythic are developing analog matrix processors that aim to deliver the compute resources of a GPU at a fraction of the power consumption.
How might analog computing change the devices we use in our daily lives?
-As analog computing becomes more integrated with digital systems, we could see devices that are always on, like voice-enabled wearables and smart home sensors, consuming much less power. This could lead to longer battery life and reduced environmental impact without sacrificing functionality.
What are some ways for individuals to explore analog computing at home?
-Individuals can explore analog computing at home through models like the Analog Paradigm Model-1, which is designed for experienced users to assemble themselves, or The Analog Thing (THAT), which is sold fully assembled and can be used for a variety of applications from simulating natural sciences to creating music.
Outlines
đș The Resurgence of Analog Computing
This paragraph introduces the concept of analog computing and its resurgence in modern technology. It discusses the shift from analog to digital computing and the potential of analog computing to impact daily life. The speaker, Matt Ferrell, shares his curiosity about analog computing sparked by a Veritasium video and his subsequent exploration of the topic. The contrast between analog and digital systems is highlighted, emphasizing the infinite states of analog versus the finite states of digital, represented by bits. The energy efficiency of analog computing is also mentioned as a potential solution to the growing energy demands of digital computing, particularly in the context of cryptocurrencies and AI.
đĄ Historical Analog Computers and Their Applications
This paragraph delves into the history and practical applications of analog computers. It mentions the Moniac National Income Analogue Computer (MONIAC) as a prime example, which was designed to simulate the British economy. The paragraph also discusses the accuracy of analog computers and their continued relevance, such as pilots using slide rules for calculations. The contrast between the convenience of digital devices and the specialized applications of analog computers is explored, highlighting the limitations of digital computing and the potential for analog computing to break through these barriers.
đ Pushing the Limits of Digital Computing
This paragraph examines the limitations of digital computing, referencing the predictions made by Gordon Moore, known as Moore's Law, and the physical boundaries that digital elements are reaching. It discusses the challenges of miniaturizing computer chips further and the heat generation and cooling requirements of dense components. The paragraph also touches on Amdahl's law and its implications for the efficiency of digital computers, especially when considering sequential operations and the diminishing returns of adding more processors. The potential of analog computing to offer a more parallel and energy-efficient approach is contrasted with the sequential nature of digital computing.
đ Future of Analog Computing in Everyday Life
The final paragraph explores the future possibilities of analog computing in everyday life, discussing the development of hybrid computers that combine the energy efficiency of analog with the precision of digital. It mentions companies like Mythic and Aspinity that are working on analog chips for machine learning and low-power sensors. The potential applications of analog computing in household devices are highlighted, such as voice-enabled wearables and heart rate monitoring. The paragraph also addresses the challenges of making analog programming more accessible and the need for seamless connectivity between analog and digital systems. It concludes with a call to action for the audience to consider the potential of analog computing and engage in further discussion.
Mindmap
Keywords
đĄAnalog computing
đĄDigital computing
đĄEnergy efficiency
đĄMoore's Law
đĄAmdahl's Law
đĄHybrid computing
đĄMachine learning
đĄData centers
đĄDifferential equations
đĄSurfshark
đĄClimate crisis
Highlights
Analog computing is making a comeback and is also something that never really left.
Analog systems have an infinite number of states, unlike digital systems which rely on a fixed number of states.
Digital computing is becoming increasingly energy intensive, with significant implications for global energy consumption.
Analog computing could be part of the solution to energy efficiency, as it can accomplish tasks for a fraction of the energy.
The MONIAC, created in 1949, is an example of an analog computer used to simulate the economy.
Pilots still use flight computers, a form of slide rule, for calculations without the need for electricity.
Digital devices provide convenience, but analog computing has its own strengths, such as energy efficiency.
Digital computers are hitting basic physical boundaries, limiting how much further they can be shrunk.
Moore's Law, which predicts the doubling of transistors on a chip, is nearing its limits.
The more components on a chip, the harder it is to cool, leading to significant energy and resource use.
Research on new approaches to analog computing has led to the development of materials that donât need cooling facilities.
Amdahl's law suggests that there will always be operations that must be performed sequentially in digital computing.
Analog computers can work in parallel, allowing for more efficient problem-solving without the need for sequential operations.
Hybrid computers that combine the best features of both digital and analog computing may be the future.
Mythic's Analog Matrix Processor chip aims to deliver significant compute resources at a fraction of the power consumption.
Aspinity's AML100 chip can act as a low-power sensor for various applications, with potential energy savings of up to 95%.
Analog computing, with its potential for energy efficiency and real-time processing, could become more approachable and accessible.
Transcripts
If your taste in TV is anything like mine, then most of your familiarity with Â
what analog computing looks like probably comes from the backdrops of something like
Columbo. Since digital took over the world, analog has been sidelined into what seems Â
like a niche interest at best. But this retro approach to computing, much like space operas, Â
is both making a comeback, and also something that never really left in the first place.
I found this out for myself about a year ago, when a video from Veritasium sparked Â
my curiosity about analog computing. After that, I started to read a few articles here and there, Â
and I gotta sayâŠit broke my brain a bit. What I really wanted to know, though, was this: Â
How can analog computing impact our daily lives? And what will that look Â
like? Because I definitely donât have room in my house for this.
Iâm Matt Ferrell ⊠welcome to Undecided.Â
This video is brought to you by Surfshark and all of my patrons on Patreon, but more on that later.
Depending on how old you are, you may remember when it was the norm for a single computer to Â
take up more square footage than your average New York City apartment. But after the end of the Â
Space Age and the advent of personal computers, our devices have only gotten smaller and smaller. Â
Some proponents of analog computing argue that we might just be reaching our limits when it Â
comes to how much further we can shrink. Weâll get to that in a bit, though. Emphasis on bits.
Speaking of bits, this brings us to the fundamental difference between analog Â
and digital. Analog systems have an infinite number of states. If I were to heat this room Â
from 68 F to 72 F, the temperature would pass through an infinite set of numbers, Â
including 68.0000001 F and so on. Digital systems are reliant on the number of âbitsâ Â
or the number of transistors that are switched either on or off. As an example, Â
an 8-bit system has 2^8, or 256 states. That means it can only represent 256 different numbers.
So, size isnât the only aspect of the technological zeitgeist thatâs changed. Digital Â
computers solve problems in a fundamentally different way from analog ones. Thatâs led Â
to some pretty amazing stuff in modern dayâŠat a cost. Immensely energy intensive computing Â
is becoming increasingly popular. Just look at cryptocurrencies and AI. According to a report Â
released last year by Swedish telecommunications company Ericsson, the information and Â
communication technology sector accounted for roughly 4% of global energy consumption in 2020.
Plus, a significant amount of digital computing is not the kind you can take to Â
go. Just among the thousands of data centers located across the globe, the average campus Â
size is approximately 100,000 square feet (or just over 9,000 square meters). That's more Â
than 2 acres of land! Data scientist Alex de Vries estimates that a single Â
interaction with a LLM is equivalent to âleaving a low-brightness LED lightbulb on for one hour.â
But as the especially power-hungry data centers, neural networks, and cryptocurrencies of the world Â
continue to grow in scale and complexityâŠwe still have to reckon with the climate crisis. Energy Â
efficiency isnât just good for the planet, itâs good for the wallet. A return to analog Â
computing could be part of the solution. The reason why is simple: you can accomplish the Â
same tasks as you would on a digital setup for a fraction of the energy. In some cases, Â
analog computing is as much as 1,000 times more efficient than its digital counterparts.
Before we get into exactly how it works and why weâre starting to see more interest in analog Â
computers again, I need to talk about another piece of tech that can really help in your daily Â
digital life and thatâs todayâs sponsor, Surfshark. Surfshark is a fast, easy to Â
use VPN full of incredible features that you can install on an unlimited number of devices with one Â
account. Most of the time when we talk about VPNs weâre focused on giving yourself security as you Â
travel around the world, but it can do way more than that. Since you can make it look like your Â
IP address is coming from somewhere else in the world, it unlocks geofencing blocks on content, Â
like streaming services. But ⊠thatâs not all. Even shopping services will sometimes Â
gate prices based on your location, so you can change your location to make sure youâre getting Â
the best prices. They also have add-ons to their VPN service to unlock things like Surfshark Alert, Â
which will let you know if your email or personal details, like passwords, have been leaked online Â
in a data breach. Right now theyâre running a special deal ⊠use my code UNDECIDED to get up Â
to 3 additional months for free. SurfShark offers a 30-day money-back guarantee, Â
so thereâs no risk to try it out for yourself. Iâve been using Surfshark for years and love it. Â
Link is in the description below. Thanks to Surfshark, for supporting the channel. Â
And thanks to all of you, as well as my patrons, who get early, ad-free versions of my videos. So Â
back to how much more energy efficient analog computing is from its digital counterparts.
To understand how that works, exactly, we first need to establish what makes analog Â
computingâŠanalog. The same way you would make a comparison with words using an analogy, Â
analog computers operate using a physical model that corresponds to the values of Â
the problem being solved. And yeah, I did just make up an analog analogy.
A classic example of analog computing is the Monetary National Income Analogue Computer, Â
or MONIAC, which sounds like a long forgotten car brand, which economist Bill Phillips created in Â
1949. MONIAC has a single purpose: to simulate the Great British economy on a macro level. Â
Within the machine, water represented money as it literally flowed in and out of the treasury. Â
Phillips determined alongside his colleague Walter Newlyn that the computer could function Â
with an approximate accuracy of ±2%. And of the 14 or so machines that were made, Â
you can still find the first churning away at the Reserve Bank Museum in New Zealand.
Itâs safe to say that the MONIAC worked (and continues to work) well. The same goes Â
for other types of analog computers, from those on the simpler end of the spectrum, Â
like the pocket-sized mechanical calculators known as slide rules, Â
to the behemoth tide-predicting machines invented by Lord Kelvin.
In general, it was never that analog computing didnât do its job â quite the opposite. Pilots Â
still use flight computers, a form of slide rule, to perform calculations by hand, Â
no juice necessary. But for more generalized applications, digital devices just provide a level Â
of convenience that analog couldnât. Incredible computing power has effectively become mundane.
To put things into perspective, an iPhone 14 contains a processor that runs somewhere Â
above 3 GHz, depending on the model. The Apollo Guidance Computer, itself a Â
digital device onboard the spacecraft that first graced the moonâs surface, Â
ran atâŠ0.043 MHz. As computer science professor Graham Kendall once wrote, Â
âthe iPhone in your pocket has over 100,000 times the processing power of the computer that landed Â
man on the moon 50 years ago.â ⊠and we use it to look at cat videos and argue with strangers.
In any case, that ease of use is one of the reasons why the likes of slide rules and Â
abacuses were relegated to museum displays while electronic calculators reigned king. Â
So much for âruling.â But, while digital has a lot to offer, like anything else, Â
it has its limits. And mathematician and self-described âanalog computer Â
evangelistâ Bernd Ulmann argues that we canât push those limits much further. In his words:
âDigital computers are hitting basic physical boundaries by now. Computing Â
elements cannot be shrunk much more than today, Â
and there is no way to spend even more energy on energy-hungry CPU chips today.â
Itâs worth noting here that Ulmann said this in 2021, years ahead of the explosion Â
of improvements in generative AI weâve witnessed in just the past few months, Â
like OpenAIâs text-prompt-to-video model, Sora. Which, really disturbs Â
me and I'm very excited by all at the same time, I need to make a video about that.
But what did he mean by âphysical boundariesâ? WellâŠdigital computing Â
is starting to bump up against the law. No, not that kindâŠthe scientific kind. Â
Thereâs actually a few that are at play here. Weâve already started talking about Â
the relationship between digital computing and size, so letâs continue down that track.
In a 1965 paper, Gordon Moore, co-founder of Intel, made a prediction that would come to Â
be known as âMooreâs Law.â He foresaw that the number of transistors on an integrated Â
circuit would double every year for the next 10 years, with a negligible rise in Â
cost. And 10 years later, Moore changed his prediction to a doubling every two years.
As Intel clarifies, Mooreâs Law isnât a scientific observation, and Moore actually isnât too keen on Â
his work being referred to as a âlaw.â However, the prediction has more or less stayed true as Â
Intel (and other semiconductor companies)Â have hailed it as a goal to strive for:Â Â
more and more transistors on smaller and smaller chips, for less and less money.
Hereâs the problem. What happens when we canât make a computer chip any smaller? According to Â
Intel, despite the warnings of experts in the past few decades, weâve yet to hit that wall. We can Â
take it straight from Moore himself, though, that an end to the standard set by his law is Â
inevitable. When asked about the longevity of his prediction during a 2005 interview, he said this:
âThe fact that materials are made of atoms is the fundamental limitation and it's not that Â
far away. You can take an electron micrograph from some of these pictures of some of these devices, Â
and you can see the individual atoms of the layers. The gate insulator in the most Â
advanced transistors is only about three molecular layers thickâŠWe're pushing up against some fairly Â
fundamental limits, so one of these days we're going to have to stop making things smaller."
Not to mention, the more components you cram onto a chip, the hotter it becomes during use, Â
and the more difficult it is to cool down. Itâs simply not possible to use all the transistors Â
on a chip simultaneously without risking a meltdown. This is also a critical problem Â
in data centers, because itâs not only electricity use that represents a huge Â
resource sink. Larger sites that use liquid as coolant rely on massive amounts of water Â
a day â think upwards of millions of gallons. In fact, Googleâs data centers in The Dalles, Â
Oregon, account for over a quarter of the cityâs water use.
Meanwhile, emerging research on new approaches to analog computing has Â
led to the development of materials that donât need cooling facilities at all.
Then thereâs another law that stymies the design of digital computers: Â
Amdahlâs law. And you might be able to get a sense of why itâs relevant just by looking at Â
your wrist. Or your wall. Analog clocks, the kind with faces, can easily show us more advantages of Â
analog computing. When the hands move forward on a clock, they do so in one continuous movement, Â
the same way analog computing occurs in real time, with mathematically continuous data. But when you Â
look at a digital clock, youâll notice that it updates its display in steps. Thatâs because, Â
unlike with analog devices, digital information is discrete. Itâs something that you count, Â
rather than measure, hence the binary format of 0s and 1s.
When a digital computer tackles a problem, it follows an algorithm, a finite number Â
of steps that eventually lead to an answer. Presenting a problem to an analog computer is Â
a completely different procedure, and this cute diagram from the â60s still holds true today:
First, you take note of the physical laws that form the context of the problem youâre Â
solving. Then, you create a differential equation that models the problem. If your Â
blood just ran cold at the mention of math, donât worry. All you need to know Â
is that differential equations model dynamic problems, or problems that involve an element Â
of change. Differential equations can be used to simulate anything from heat flow in a cable Â
to the progression of zombie apocalypses. And analog computers are fantastic at solving them.
Once youâve written a differential equation, you program the analog computer by translating Â
each part of the equation into a physical part of the computer setup. And then you get your answer, Â
which doesnât even necessarily require a monitor to display!
All of that might be tough to envision, so hereâs another analog analogy that hopefully Â
is less convoluted than the labyrinth of wires that make up a patch panel. Imagine a playground. Â
Letâs say two kids want to race to the same spot, but each one takes a different path. Â
One decides to skip along the hopscotch court, and the other rushes to the slide. Who will win?
These two areas of the playground are like different paradigms of computing. Â
You count the hopscotch spaces outlined on the ground and move between them one by one, Â
but you measure the length of a slide, and reach its end in one Â
smooth move. And between these two methods of reaching the same goal, Â
one is definitely a much quicker process than the otherâŠand also takes a lot less energy.
There are, of course, caveats to analog. If you asked the children in our playground Â
example to repeat their race exactly the same way they did the first time, who do you think Â
would be more accurate? Probably the one whose careful steps were marked with neat squares, Â
and whose outcomes will be the same â landing right within that final little perimeter of Â
chalk. With discrete data, you can make perfect copies. Itâs much harder to create copies with Â
the more messy nature of continuous data. The question is: do we even need 100% accurate Â
calculations? Some researchers are proposing that we donât, at least not all the time.
That said, what does this have to do with Amdahlâs law? Well, we can extend our existing scenario Â
a little further. It takes time to remember the rules of hopscotch and then follow them Â
accordingly. But you donât need to remember any rules to use a slide â other than maybe âwait Â
until there isnât anybody else on it.â Comment below with your favorite playground accidents!
In any case, because digital computers 1. reference their memories and 2. solve problems Â
algorithmically, there will always be operations (like remembering hopscotch rules) that must be Â
performed sequentially. As computer science professor Mike Bailey puts it, âthis includes Â
reading data, setting up calculations, control logic, storing results, etc.â And because you Â
canât get rid of these sequential operations, you run into diminishing returns as you add Â
more and more processors in attempts to speed up your computing. You canât decrease the size of Â
components forever, and you canât increase the number of processors forever, either.
On the other hand, analog computers donât typically have memories they Â
need to take time to access. This allows them more flexibility to work in parallel, Â
meaning they can easily break down problems into smaller, Â
more manageable chunks and divide them between processing units without delays.
Hereâs how Bernd Ulmann explains it In his 2023 textbook, Analog and Hybrid Computer Programming, Â
which contributed a considerable amount of research to this video:
âFurther, without any memory there is nothing like a critical section, Â
no need to synchronize things, no communications overhead, Â
nothing of the many trifles that haunt traditional parallel digital computers.â
So, you might be thinking: speedier, more energy-efficient computing sounds great, Â
but what does it have to do with me? Am I going to have to learn how to write differential equations? Â
Will I need to knock down a wall in my office to make room for a retrofuturist analog rig?
Probably not. Instead, hybrid computers that marry the best features of both digital and Â
analog are what might someday be in vogue. Thereâs already whisperings of Silicon Valley Â
companies secretly chipping away atâŠanalog chips. Why? To conserve electricity ⊠and Â
cost. The idea is to combine the energy efficiency of analog with the precision of Â
digital. This is especially important for continued development of the power-hungry Â
machine learning that makes generative AI possible. With any hope, that means Â
products that are far less environmentally and financially costly, to maintain.
And thatâs exactly what Mythic, headquartered in the U.S., is aiming for. Mythic claims that its Â
Analog Matrix Processor chip can âdeliver the compute resources of a GPU at 1/10th the power Â
consumption.â Basically, as opposed to storing data in static RAM, which needs an uninterrupted Â
supply of power, the analog chip stores data in flash memory, which doesnât need power to Â
keep information intact. Rather than 1s and 0s, the data is retained in the form of voltages.
Where could we someday see analog computing around the house, though? U.S.-based company Â
Aspinity has an answer to that. What it calls the âworldâs first fully analog Â
machine learning chip,â the AML100, can act as a low-power sensor for a bunch of applications, Â
according to its website. It can detect a wake word for use in voice-enabled wearables like Â
wireless earbuds or smart watch, listen for the sound of broken glass or smoke alarms, Â
and monitor heart rates, just to name a few.
For those devices that always need to be on, this means energy savings that are Â
nothing to sneeze at (although I guess you could program an AML 100 to detect Â
sneezes). Aspinity claims that its chip can enable a reduction in power use of 95%.
So, the potential of maximizing efficiency through analog computing is clear, Â
and the world we interact with every day is itself analog. Why shouldnât our devices be, too? But to Â
say that analog programming appears intimidating (and dated) isâŠsomewhat of an understatement.
Itâll definitely need an image upgrade to make it approachable and accessible to Â
the public â though there are already models out there that you can fiddle with yourself at home, Â
if youâre brave enough. German company Anabrid, which was founded by Ulmann in 2020, Â
currently offers two: the Analog Paradigm Model-1, and The Analog Thing (or THAT).
The Model-1 is intended for more experienced users who are willing to assemble the machine Â
themselves. Each one is produced on demand based on the parts ordered, Â
so you can tailor the modules to your needs.
THAT, on the other handâŠand by THAT I mean THAT: The Analog Thing, is sold fully assembled. You Â
could also build your own from scratch â the components and schematics are open source.
So what do you actually do with the thing? YâknowâŠTHAT? Â
Iâll let the official wikiâs FAQ answer that:
âYou can use it to predict in the natural sciences, to control in engineering, Â
to explain in educational settings, to imitate in gaming, or you can use it for the pure joy of it.â
The THAT model, like any analog computer, solves whatever you can express in a Â
differential equation. As a reminder, thatâs basically any scenario involving change, Â
from simulating air flow to solving heat equations. You can also make music!
But as analog computing becomes more readily available, thereâs still a Â
lot of work to be done. For one thing, Itâll take effort to engineer seamless Â
connectivity between analog and digital systems, as Ulmann himself points out.
Until then, what do you think? Should we take the word of analog evangelists as gospel? Or Â
are we better off waiting for flying cars? Jump into the comments and let me know. Be Â
sure to check out my follow-up podcast, Still To Be Determined, where we'll be discussing Â
some of your feedback. Before I go, Iâd like to welcome new Supporter+ patrons Â
Charles Bevitt and Tanner. Thanks so much for your support. Iâll see you in the next one.
5.0 / 5 (0 votes)