So far we’ve covered the sensors/devices that are out in the world collecting data, and the connectivity technologies that enable those sensors/devices to pass that data up to the cloud for processing. But what is the cloud? And what happens when that data is received?
Back in the 1970s, it was popular for businesses to rent time using big, mainframe computer systems. These systems were extremely large and expensive, so it didn’t make sense financially for businesses to own the computing power themselves. Instead, they were owned by large corporations, government agencies, and universities.
Microprocessor technology allowed for great reductions in size and expense, leading to the advent of the personal computer, which exploded in popularity in the 1980s. Suddenly, businesses could (and did) bring computation in-house.
However, as high-speed connections have become widespread, the trend has reversed: businesses are once again renting computing power from other organizations. But why is that?
Instead of buying expensive hardware for storage and processing in-house, it’s easy to rent it for cheap in the cloud. The cloud is a huge, interconnected network of powerful servers that performs services for businesses and for people.
The largest cloud providers are Amazon, Google, and Microsoft, who have huge farms of servers that they rent to businesses as part of their cloud services.
For businesses that have variable needs (most of the time they don’t need much computing, but every now and then they need a lot), this is cost effective because they can simply pay as-needed.
When it comes to people, we use these cloud services all of the time. You might store your files in Google Drive instead of on your personal computer. Google Drive, of course, uses Google’s cloud services.
Or you might listen to songs on Spotify instead of downloading the songs to your computer or phone. Spotify uses Amazon’s cloud services.
Generally, something that happens “in the Cloud” is any activity that takes place over an internet connection instead of on the device itself.
Because activities like storage and data processing take place in the cloud rather than on the device itself, this has had significant implications for IoT.
Many IoT systems make use of large numbers of sensors to collect data and then make intelligent decisions.
Using the cloud is important for aggregating data and drawing insights from that data. For instance, a smart agriculture company would be able to compare soil moisture sensors from Kansas and Colorado after planting the same seeds. Without the cloud, comparing data across wider areas is much more difficult.
Using the cloud also allows for high scalability. When you have hundreds, thousands, or even millions of sensors/devices, putting large amounts of computational power on each sensor/device would be extremely expensive and energy intensive. Instead, data can be passed to the cloud from all these sensors and processed there in aggregate.
For much of IoT, the head (or rather, the brain) of the system is in the cloud. Sensors/devices collect data and perform actions, but the processing/commanding/analytics (aka the “smart” stuff), typically happens in the cloud.
Technically, the answer is no. The data processing and commanding could take place locally rather than in the cloud via an internet connection. Known as “fog computing” or “edge computing”, this actually makes a lot of sense for some IoT applications.
However, there are substantial benefits to be had using the cloud for many IoT applications including:
There are legitimate concerns with cloud usage though:
The Internet of Things is a broad field and includes an incredible variety of applications. There is no one-size-fits-all solution so you need to consider your organization’s specific application when deciding whether the cloud makes sense.