Toggle light / dark theme

The Metaverse is Taking Over the Physical World

Imagine a place where you could always stay young, name a city after yourself, or even become the president — sounds like a dream? Well, if not in the real world, such dreams can definitely be fulfilled in the virtual world of a metaverse. The metaverse is believed by some to be the future of the internet, where apart from surfing, people would also be able to enter inside the digital world of the internet, in the form of their avatars.

The advent of AR, blockchain, and VR devices in the last few years has sparked the development of the metaverse. Moreover, the unprecedented growth of highly advanced technologies in the gaming industry, which offer immersive gameplay experiences, not only provides us a glimpse of how the metaverse would look like but also indicates that we are closer than ever to experience a virtual world of our own.

Building a template for the future 6G network

Traditional networks are unable to keep up with the demands of modern computing, such as cutting-edge computation and bandwidth-demanding services like video analytics and cybersecurity. In recent years, there has been a major shift in the focus of network research towards software-defined networks (SDN) and network function virtualization (NFV), two concepts that could overcome the limitations of traditional networking. SDN is an approach to network architecture that allows the network to be controlled using software applications, whereas NFV seeks to move functions like firewalls and encryption to virtual servers. SDN and NFV can help enterprises perform more efficiently and reduce costs. Needless to say, a combination of the two would be far more powerful than either one alone.

In a recent study published in IEEE Transactions on Cloud Computing, researchers from Korea now propose such a combined SDN/NFV network architecture that seeks to introduce additional computational functions to existing network functions. “We expect our SDN/NFV-based infrastructure to be considered for the future 6G network. Once 6G is commercialized, the resource management technique of the network and computing core can be applied to AR/VR or holographic services,” says Prof. Jeongho Kwak of Daegu Gyeongbuk Institute of Science and Technology (DGIST), Korea, who was an integral part of the study.

The new network architecture aims to create a holistic framework that can fine-tune processing resources that use different (heterogeneous) processors for different tasks and optimize networking. The unified framework will support dynamic service chaining, which allows a single network connection to be used for many connected services like firewalls and intrusion protection; and code offloading, which involves shifting intensive computational tasks to a resource-rich remote server.

This AI Makes Digital Copies of Humans! 👤

❤️ Check out the Gradient Dissent podcast by Weights & Biases: http://wandb.me/gd.

📝 The paper “The Relightables: Volumetric Performance Capture of Humans with Realistic Relighting” is available here:
https://augmentedperception.github.io/therelightables/

🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
Aleksandr Mashrabov, Alex Haro, Andrew Melnychuk, Angelos Evripiotis, Benji Rabhan, Bryan Learn, Christian Ahlin, Eric Haddad, Eric Martel, Gordon Child, Ivo Galic, Jace O’Brien, Javier Bustamante, John Le, Jonas, Kenneth Davis, Klaus Busse, Lorin Atzberger, Lukas Biewald, Matthew Allen Fisher, Mark Oates, Michael Albrecht, Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Ramsey Elbasheer, Steef, Taras Bobrovytsky, Thomas Krcmar, Timothy Sum Hon Mun, Torsten Reil, Tybie Fitzhugh, Ueli Gallizzi.
If you wish to appear here or pick up other perks, click here: https://www.patreon.com/TwoMinutePapers.

Thumbnail background design: Felícia Zsolnai-Fehér — http://felicia.hu.

Károly Zsolnai-Fehér’s links:
Instagram: https://www.instagram.com/twominutepapers/
Twitter: https://twitter.com/twominutepapers.
Web: https://cg.tuwien.ac.at/~zsolnai/

#vr

RocketVirtual Blog

I finally completed my extensive WebXR blog and multi-user VR chat server examples at these two link address… You are welcome to share the knowledge with others interested in creating a complete VR experience inside the browser using JavaScript libraries and HTML5. Check it out, enjoy, and follow me on twitter for recent updates. https://funbit64.com/ #technology #javascript #webxr #virtualreality #oculusquest2 #metaverse #webdeveloper #blog #sourcecode #learncoding #share #html #aframe #vrtraining #vrtechnology #vrexperience #vrdevelopment #vrheadset #vrgame #immersive #immersivetechnology #immersiveexperiences


A virtual reality blog for VR software developers, and VR content creators. Current and relevant working examples, original content and blogger commentary. Making use of VR technologies: WebXR, A-Frame, WebVR (depricated), JavaScript and HTML5.

Jaron Lanier — Is Consciousness an Ultimate Fact?

Is there something special about consciousness? Can our inner subjective experience—the sight of purple, smell of cheese, sound of Bach—ever be explained in purely physical terms? Even in principle? Most scientists see consciousness as a science problem to solve. Some philosophers claim that consciousness can never be explained in terms of current science.

Free access to Closer to Truth’s library of 5,000 videos: http://bit.ly/376lkKN

Watch more interviews on consciousness: https://bit.ly/3Ce962v.

Jaron Zepel Lanier is an American computer scientist, best known for popularizing the term virtual reality (VR).

Register for free at CTT.com for subscriber-only exclusives: http://bit.ly/2GXmFsP

Closer to Truth, hosted by Robert Lawrence Kuhn, presents the world’s greatest thinkers exploring humanity’s deepest questions. Discover fundamental issues of existence. Engage new and diverse ways of thinking. Appreciate intense debates. Share your own opinions. Seek your own answers.

US Navy, Boeing conduct first-ever refueling between unmanned tanker, F-35C

“Once operational, the MQ-25 will refuel every receiver-capable platform including E-2,” Reed said in a separate news release on the August flight test between T1 and the Navy’s Air Test and Evaluation Squadron Two Zero (VX) 20. “This flight keeps us on a fast track to getting the Stingray out to the fleet where its refueling capability will greatly increase the range and operational flexibility of the carrier air wing and strike group.”

The first aerial refueling test was conducted on June 4, when an F/A-18E-F Super Hornet refueled in air with the unmanned tanker for the first time in naval aviation history. In this first in-air encounter with the drone, a Super Hornet approached to take measurements, made several “dry connects” to practice connecting and detaching from the tanker, and made two actual refuelings, with 300 pounds and then 25 pounds of fuel being passed from the Stingray to the Super Hornet at different altitudes and flying conditions.

Last month, the Navy and Boeing conducted in a virtual environment the first manned-unmanned teaming event between the Stingray and a Super Hornet, where the manned jet bypassed the ground control station and communicated directly with the tanker drone to give directions on where and when to rendezvous for a refueling.

New Chip Can Decode Any Type of Data Sent Across a Network

Every piece of data that travels over the internet — from paragraphs in an email to 3D graphics in a virtual reality environment — can be altered by the noise it encounters along the way, such as electromagnetic interference from a microwave or Bluetooth device. The data are coded so that when they arrive at their destination, a decoding algorithm can undo the negative effects of that noise and retrieve the original data.

Since the 1950s, most error-correcting codes and decoding algorithms have been designed together. Each code had a structure that corresponded with a particular, highly complex decoding algorithm, which often required the use of dedicated hardware.

Researchers at MIT.

/* */