> Edge computing is a paradigm that sees complex computational tasks – like data analysis or cloud gaming – performed as close to the “user” as possible. By running these tasks closer to the source or consumer, rather than in a vast air-cooled Amazon data centre, you reduce latency.
So basically it's a hip and trendy new name for the way computing has been done for the last half century?