home.social

#realitycollective — Public Fediverse posts

Live and recent posts from across the Fediverse tagged #realitycollective, aggregated by home.social.

  1. Sometimes you need to know whether or not the user is now actually wearing the device while your app is running. I wrote a little #RealityCollective #ServiceFramework Service to make that easily detectable #MixedReality #HoloLens2

    localjoost.github.io/Detecting

  2. Sometimes you need to know whether or not the user is now actually wearing the device while your app is running. I wrote a little Service to make that easily detectable

    localjoost.github.io/Detecting

  3. Sometimes you need to know whether or not the user is now actually wearing the device while your app is running. I wrote a little #RealityCollective #ServiceFramework Service to make that easily detectable #MixedReality #HoloLens2

    localjoost.github.io/Detecting

  4. Sometimes you need to know whether or not the user is now actually wearing the device while your app is running. I wrote a little #RealityCollective #ServiceFramework Service to make that easily detectable #MixedReality #HoloLens2

    localjoost.github.io/Detecting

  5. Sometimes you need to know whether or not the user is now actually wearing the device while your app is running. I wrote a little #RealityCollective #ServiceFramework Service to make that easily detectable #MixedReality #HoloLens2

    localjoost.github.io/Detecting

  6. For my second blogpost of the day ;) one of my more infamous and far-fetched experiments encompasses the combination of #MachineLearning combined with a Spatial Map to recognize and locate objects in 3D space, using a #YoloV8 #computervision model. This worked pretty well on #HoloLens2. I got it to work on #MagicLeap2 as well, but it was quite a rocky road. The #RealityCollective #ServiceFramework to the rescue :). Explanation, sample and full code at: localjoost.github.io/Running-a

  7. For my second blogpost of the day ;) one of my more infamous and far-fetched experiments encompasses the combination of combined with a Spatial Map to recognize and locate objects in 3D space, using a model. This worked pretty well on . I got it to work on as well, but it was quite a rocky road. The to the rescue :). Explanation, sample and full code at: localjoost.github.io/Running-a

  8. For my second blogpost of the day ;) one of my more infamous and far-fetched experiments encompasses the combination of #MachineLearning combined with a Spatial Map to recognize and locate objects in 3D space, using a #YoloV8 #computervision model. This worked pretty well on #HoloLens2. I got it to work on #MagicLeap2 as well, but it was quite a rocky road. The #RealityCollective #ServiceFramework to the rescue :). Explanation, sample and full code at: localjoost.github.io/Running-a

  9. For my second blogpost of the day ;) one of my more infamous and far-fetched experiments encompasses the combination of #MachineLearning combined with a Spatial Map to recognize and locate objects in 3D space, using a #YoloV8 #computervision model. This worked pretty well on #HoloLens2. I got it to work on #MagicLeap2 as well, but it was quite a rocky road. The #RealityCollective #ServiceFramework to the rescue :). Explanation, sample and full code at: localjoost.github.io/Running-a

  10. For my second blogpost of the day ;) one of my more infamous and far-fetched experiments encompasses the combination of #MachineLearning combined with a Spatial Map to recognize and locate objects in 3D space, using a #YoloV8 #computervision model. This worked pretty well on #HoloLens2. I got it to work on #MagicLeap2 as well, but it was quite a rocky road. The #RealityCollective #ServiceFramework to the rescue :). Explanation, sample and full code at: localjoost.github.io/Running-a