Straight out of science fiction! Drones that can track and 3D reconstruct any person also while avoiding obstacles! (pose estimation)

paper link

https://reddit.com/link/qhbrdm/video/usqhai2og3w71/player

---

For more exciting developments like this, follow CatalyzeX on twitter, and add our browser extension to easily get code for machine learning / AI papers anywhere!πŸ‘‡

Chrome: https://chrome.google.com/webstore/detail/aiml-papers-with-code-eve/aikkeehnlfpamidigaffhfmgbkdeheil

Firefox: https://addons.mozilla.org/en-US/firefox/addon/code-finder-catalyzex

πŸ‘︎ 21
πŸ’¬︎
πŸ‘€︎ u/fullerhouse570
πŸ“…︎ Oct 28 2021
🚨︎ report
Making a 3D model move using Tensorflow Pose Estimation and Three.js! youtu.be/Vvo0LfjcLNU
πŸ‘︎ 19
πŸ’¬︎
πŸ‘€︎ u/AphrxWasHere
πŸ“…︎ Oct 20 2021
🚨︎ report
Straight out of science fiction! Drones that can track and 3D reconstruct any person also while avoiding obstacles! (pose estimation) /r/LatestInML/comments/qh…
πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/fullerhouse570
πŸ“…︎ Oct 28 2021
🚨︎ report
Straight out of science fiction! Drones that can track and 3D reconstruct any person also while avoiding obstacles! (pose estimation) /r/LatestInML/comments/qh…
πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/fullerhouse570
πŸ“…︎ Oct 28 2021
🚨︎ report
Straight out of science fiction! Drones that can track and 3D reconstruct any person also while avoiding obstacles! (pose estimation) /r/LatestInML/comments/qh…
πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/fullerhouse570
πŸ“…︎ Oct 28 2021
🚨︎ report
labelCloud for 3D Object Detection and 6D Pose Estimation

Hi everyone! :)

One of my former students developed a very cool point cloud labeling tool, called labelCloud! The tool provides smooth labeling interaction with 3D point clouds to support users in generating training data for #3DObjectDetection and #6DPoseEstimation.

labelCloud is publicly available, easy to install, and simple to use. Check out our github: https://github.com/ch-sa/labelCloud

If you give it a go, we would be more than happy to receive your feedback on it. So, as we are currently evaluating it, we invite you to fill this short questionaire https://forms.gle/moEyjGSa1Eiiq7VT8 (~5 min)! Thanks in advance! :)

Further information can be found in our paper from CAD'21 conference: http://dx.doi.org/10.14733/cadconfP.2021.319-323

πŸ‘︎ 15
πŸ’¬︎
πŸ‘€︎ u/patzscheck
πŸ“…︎ Aug 27 2021
🚨︎ report
3D Pose Estimation from 2D image and its bounding box

Hello guys!

I am working on a problem that requires estimating the 3D rotation and translation of an object with respect to the reference camera. Please find below details on the expected input and output of the pose estimation algorithm.

Can anyone please help me out and guide me on how I can achieve this??

https://preview.redd.it/jduifqo8s6d71.png?width=581&format=png&auto=webp&s=9d98bdf1c4b9919c3211151973653345d0dc4364

Expected Input:

  • The input to the pose estimation algorithm can be an image or a sequence of images depending on the logic.
  • The bounding box from object detection can be used to localize the object in the 2D image.

Solution/Output Expected:

The output expected is divided into 2 categories:

  • Part 1: Provides the depth of the centroid of the object. This corresponds to the Z-axis coordinates.
  • Part 2: Provides the 6DOF pose from the pose estimation algorithm i.e. the position of the object with respect to the camera (X, Y, and Z coordinates) and also the rotation of the object about all the 3 reference axis (yaw, pitch, and roll).

I had previously built an object detection system using the YOLOv5 algorithm which was then trained to detect objects on few custom categories. I am now wondering how to proceed further for this pose estimation problem with the bounding boxes that I obtained using YOLOv5.

πŸ‘︎ 4
πŸ’¬︎
πŸ‘€︎ u/_cryptoray_
πŸ“…︎ Jul 24 2021
🚨︎ report
Facebook Research | Monocular 3D Pose Estimation Outputs for Body, Hand, Body+Hands in a Single System v.redd.it/76hcc8m83jp51
πŸ‘︎ 86
πŸ’¬︎
πŸ‘€︎ u/AR_MR_XR
πŸ“…︎ Sep 26 2020
🚨︎ report
RULA / REBA dataset for 3D pose estimation

I am currently working on my master's thesis in 3D Pose Estimation, Computer Vision.

Therefore I am searching for a Rapid Upper Limb Assessment (RULA) or Rapid Entire Body Assessment (REBA) video dataset for 3D pose estimation with ground truth (either as 3D pose or depth). Perfect was RGB-D [d=depth] videos with GT as 3D pose.

Sample: RULA Employee Assessment Worksheet by ErgoPlus

In particular I am looking for a single-person video dataset. Ideally the person moves only one limb at a time, recorded like in a natural doctor's office (no need for articial scenery as well as non-essentially clean "laboratory conditions").

  • I know about AMASS (and alike datasets), but these require to model and animate the pose I want to see - and without MoCap that might result in unnaturally movement.
  • Further most RGB-D datasets like RGB-D People Dataset dont fit due to multi-person videos.

Feel free to even point my nose on something. Even by googleing I feel stuck now.

Thank you in advance for your time :)

I am writing my thesis in Germany where afaik RULA apps do not exist yet. I have no commercial affiliation.

πŸ‘︎ 7
πŸ’¬︎
πŸ‘€︎ u/TheVadammt
πŸ“…︎ Dec 12 2020
🚨︎ report
Facebook | State of the Art 3D Pose Estimation Outputs for Body, Hand, Body+Hands in a Single System | FrankMocap v.redd.it/76hcc8m83jp51
πŸ‘︎ 71
πŸ’¬︎
πŸ‘€︎ u/AR_MR_XR
πŸ“…︎ Sep 26 2020
🚨︎ report
[R] 3D Human Pose Estimation with Spatial and Temporal Transformers

We present PoseFormer, a purely transformer-based approach for 3D human pose estimation in videos without convolutional architectures involved.

We design a spatial-temporal transformer structure to comprehensively model the human joint relations within each frame as well as the temporal correlations across frames, then output an accurate 3D human pose of the center frame.

We quantitatively and qualitatively evaluate our method on two popular and standard benchmark datasets: Human3.6M and MPI-INF-3DHP.
Extensive experiments show that PoseFormer achieves state-of-the-art performance on both datasets.

Paper: https://arxiv.org/pdf/2103.10455.pdf
Code: https://github.com/zczcwh/PoseFormer

PoseFormer: 3D Human Pose Estimation with Spatial and Temporal Transformers

πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/Extension-Sun1816
πŸ“…︎ Mar 29 2021
🚨︎ report
6D pose estimation of a known 3D CAD object

Hello, I'm working on a project where I need to estimate the 6DOF pose of a known 3D CAD object in a single RGB image - i.e. this task: https://paperswithcode.com/task/6d-pose-estimation. There are several constraints on the problem:

- Usable commercially (licensed under BSD, MIT, BOOST, etc.), not GPL.

- The CAD object is known and we do NOT aim for generality (i.e.recognize the class of all chairs).

- The CAD object can be uploaded by a user, so it may have symmetries and a range of textures.

- Inference step will be run on a smartphone, and should be able to run at >30fps.

- Can be anywhere on the scale of single instance of a single object to multiple instances of multiple objects (MiMo). MiMO is preferred, but not required.

- If a deep learning approach is used, the training time required for a new CAD object should be on the order of hours, not days.

- Can either 1) just find the initial pose of an object and not have any refinement steps after or 2) find the initial pose of the object and also have refinement steps after.

I am open to traditional approaches (i.e. 2D->3D correspondences then solving with PnP), but it seems like deep learning approaches outperform them (classical are too slow - https://stackoverflow.com/questions/62187435/real-time-6d-pose-estimation-of-known-3d-cad-objects-from-a-single-2d-image-or-p). Looking at deep learning approaches (poseCNN, HybridPose, Pix2Pose, CosyPose), it seems most of them match these constraints, except that they require model training time. Though perhaps I can use a single pre-trained model and then specialize it for each new CAD object with a shorter training step. So, my question: would somebody know of a commercially usable implementation that doesn't require extensive training time for a new CAD object?

πŸ‘︎ 14
πŸ’¬︎
πŸ‘€︎ u/gold_twister
πŸ“…︎ Sep 28 2020
🚨︎ report
The University of Notre Dame and Facebook AI Research Propose β€œImg2pose”, A Real-Time 6DoF 3D Face Pose Estimation Without Face Detection Or Landmark Localization

A new study by Facebook AI and the University of Notre Dame research team has proposed a novel real-time six DoF (Degrees of Freedom)Β 3D face pose estimation technique, namedΒ Img2pose, that works without face detection or landmark localization.

6 DoFΒ means the freedom of movement of a body in 3D space in six different ways. Other than yaw, pitch and roll rotational motion that is already there in 3 DoF, 6 DoF face pose estimation adds front/back, up/down, and left/right variables. The proposed technique can directly estimate the 6DoF 3D face pose for all faces, even in very crowded images, without the face detection step.

Summary: https://www.marktechpost.com/2020/12/23/the-university-of-notre-dame-and-facebook-ai-research-propose-img2pose-a-real-time-6dof-3d-face-pose-estimation-without-face-detection-or-landmark-localization/

Paper:Β https://arxiv.org/pdf/2012.07791.pdf

GitHub:Β https://github.com/vitoralbiero/img2pose

πŸ‘︎ 9
πŸ’¬︎
πŸ‘€︎ u/ai-lover
πŸ“…︎ Dec 23 2020
🚨︎ report
Our new 3D interacting hand pose estimation dataset (InterHand2.6M)

InterHand2.6M (ECCV 2020) is our new 3D interacting hand pose dataset.

This is the first large-scale, real-captured, and marker-less 3D interacting hand pose dataset with accurate GT 3D poses.

Checkout our InterHand2.6M

* arxiv: https://arxiv.org/abs/2008.09309

* code: https://github.com/facebookresearch/InterHand2.6M

* dataset: https://mks0601.github.io/InterHand2.6M/

* youtube: https://www.youtube.com/watch?v=h66jFalMpDQ

https://i.redd.it/5w4pqavusui51.gif

πŸ‘︎ 46
πŸ’¬︎
πŸ‘€︎ u/mks0601
πŸ“…︎ Aug 24 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image (with code available). It's called it I2L-MeshNet and here's a video I made introducing it and showing some results! youtube.com/watch?v=tDz2w…
πŸ‘︎ 63
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 54
πŸ’¬︎
πŸ“…︎ Aug 23 2020
🚨︎ report
2D 3D correspondence for Pose estimation

I would like to estimate the pose based on 2D-3D correspondence. I have tried pnp options within OpenCv. The pose is obtained by making using of the sift keypoints and the corresponding 3d points. However the estimated pose fluctuates and 50-70 cm off. Is there any other alternatives for the same for accurate pose estimation?

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/Nofapmotivation8
πŸ“…︎ Oct 28 2020
🚨︎ report
2D-3D Pose Estimation and Action Recognition using Multitask Deep Learning Code: Given in the Comment.. v.redd.it/2qw7vajm8f251
πŸ‘︎ 67
πŸ’¬︎
πŸ‘€︎ u/TheInsaneApp
πŸ“…︎ Jun 02 2020
🚨︎ report
Accurate 3D Human Pose and Mesh Estimation from a Single RGB Image!

For project and code/API/expert requests: click here

https://i.redd.it/am22n2x5yhg51.gif

πŸ‘︎ 28
πŸ’¬︎
πŸ‘€︎ u/MLtinkerer
πŸ“…︎ Aug 12 2020
🚨︎ report
[News] Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 37
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 24
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 18
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Back-Hand-Pose: 3D Hand Pose Estimation for a Wrist-worn Camera via Dorsum Deformation Network [ACM UIST2020] youtube.com/watch?v=7IJUs…
πŸ‘︎ 4
πŸ’¬︎
πŸ‘€︎ u/tcboy88
πŸ“…︎ Oct 22 2020
🚨︎ report
Synthetic Training for Accurate 3D Human Pose and Shape Estimation in the Wild

Github Paper

Most recent single-image 3D pose and shape estimation methods output pretty accurate 3D poses but inaccurate shapes, particularly for people with "non-average" body shapes. This work makes some progress towards more accurate shape estimation from a single image (although it does not work very well with baggy and loose clothing).

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/synonymous1964
πŸ“…︎ Dec 04 2020
🚨︎ report
Back-Hand-Pose: 3D Hand Pose Estimation for a Wrist-worn Camera via Dorsum Deformation Network [ACM UIST2020] youtube.com/watch?v=7IJUs…
πŸ‘︎ 5
πŸ’¬︎
πŸ‘€︎ u/tcboy88
πŸ“…︎ Oct 22 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 7
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Back-Hand-Pose: 3D Hand Pose Estimation for a Wrist-worn Camera via Dorsum Deformation Network [ACM UIST2020] youtube.com/watch?v=7IJUs…
πŸ‘︎ 4
πŸ’¬︎
πŸ‘€︎ u/tcboy88
πŸ“…︎ Oct 22 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 13
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
3D hand pose estimation using a wrist-worn camera sciencedaily.com/releases…
πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/Buddistmonk1234
πŸ“…︎ Oct 22 2020
🚨︎ report
3D hand pose estimation using a wrist-worn camera titech.ac.jp/english/news…
πŸ‘︎ 3
πŸ’¬︎
πŸ‘€︎ u/AR_MR_XR
πŸ“…︎ Nov 03 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image! (with code available) youtube.com/watch?v=tDz2w…
πŸ‘︎ 21
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Latest from UPenn researchers: Multi-person 3D pose estimation from a single image /r/LatestInML/comments/ha…
πŸ‘︎ 12
πŸ’¬︎
πŸ‘€︎ u/MLtinkerer
πŸ“…︎ Jun 17 2020
🚨︎ report
Here's a new paper announced in the ECCV2020 where they proposed a new technique for 3D Human Pose and Mesh Estimation from a single RGB image (with code available). It's called it I2L-MeshNet and here's a video I made introducing it and showing some results! youtube.com/watch?v=tDz2w…
πŸ‘︎ 3
πŸ’¬︎
πŸ“…︎ Aug 22 2020
🚨︎ report
Accurate 3D Human Pose and Mesh Estimation from a Single RGB Image! /r/LatestInML/comments/i8…
πŸ‘︎ 8
πŸ’¬︎
πŸ‘€︎ u/MLtinkerer
πŸ“…︎ Aug 12 2020
🚨︎ report
Latest from UPenn researchers: Multi-person 3D pose estimation from a single image /r/LatestInML/comments/ha…
πŸ‘︎ 50
πŸ’¬︎
πŸ‘€︎ u/MLtinkerer
πŸ“…︎ Jun 17 2020
🚨︎ report
Latest from UPenn researchers: Multi-person 3D pose estimation from a single image

For project and code/API/expert requests: click here

https://reddit.com/link/hak4jd/video/hzhyncvdfe551/player

Experiments show that their approach outperforms previous methods on standard 3D pose benchmarks, while their proposed losses enable more coherent reconstruction in natural images

πŸ‘︎ 36
πŸ’¬︎
πŸ‘€︎ u/MLtinkerer
πŸ“…︎ Jun 17 2020
🚨︎ report
[R] labelCloud for 3D Object Detection and 6D Pose Estimation

Hi everyone! :)

One of my former students developed a very cool point cloud labeling tool, called labelCloud! The tool provides smooth labeling interaction with 3D point clouds to support users in generating training data for #3DObjectDetection and #6DPoseEstimation.

labelCloud is publicly available, easy to install, and simple to use. Check out our github: https://github.com/ch-sa/labelCloud

If you give it a go, we would be more than happy to receive your feedback on it. So, as we are currently evaluating it, we invite you to fill this short questionaire https://forms.gle/moEyjGSa1Eiiq7VT8 (~5 min)! Thanks in advance! :)

Further information can be found in our paper from CAD'21 conference: http://dx.doi.org/10.14733/cadconfP.2021.319-323

πŸ‘︎ 8
πŸ’¬︎
πŸ‘€︎ u/patzscheck
πŸ“…︎ Aug 27 2021
🚨︎ report
[R] 3D Human Pose Estimation with Spatial and Temporal Transformers

We present PoseFormer, a purely transformer-based approach for 3D human pose estimation in videos without convolutional architectures involved.

We design a spatial-temporal transformer structure to comprehensively model the human joint relations within each frame as well as the temporal correlations across frames, then output an accurate 3D human pose of the center frame.

We quantitatively and qualitatively evaluate our method on two popular and standard benchmark datasets: Human3.6M and MPI-INF-3DHP.
Extensive experiments show that PoseFormer achieves state-of-the-art performance on both datasets.

Paper: https://arxiv.org/pdf/2103.10455.pdf
Code: https://github.com/zczcwh/PoseFormer

3D Human Pose Estimation on Videos in the Wild using PoseFormer

πŸ‘︎ 2
πŸ’¬︎
πŸ‘€︎ u/Extension-Sun1816
πŸ“…︎ Mar 29 2021
🚨︎ report
[R] Our new 3D interacting hand pose estimation dataset (InterHand2.6M)

InterHand2.6M (ECCV 2020) is our new 3D interacting hand pose dataset.

This is the first large-scale, real-captured, and marker-less 3D interacting hand pose dataset with accurate GT 3D poses.

Checkout our InterHand2.6M

* arxiv: https://arxiv.org/abs/2008.09309

* code: https://github.com/facebookresearch/InterHand2.6M

* dataset: https://mks0601.github.io/InterHand2.6M/

* youtube: https://www.youtube.com/watch?v=h66jFalMpDQ

https://i.redd.it/snr00rm2sui51.gif

πŸ‘︎ 9
πŸ’¬︎
πŸ‘€︎ u/mks0601
πŸ“…︎ Aug 24 2020
🚨︎ report

Please note that this site uses cookies to personalise content and adverts, to provide social media features, and to analyse web traffic. Click here for more information.