About
Tech Lead, Research Engineer at Facebook AI Research (FAIR) with 10+ years of industry…
Articles by Oleksandr
Activity
-
Just six weeks in Meta working on Llama’s multimodal capabilities, and it’s been all about moving fast. I have already contributed to Llama4’s image…
Just six weeks in Meta working on Llama’s multimodal capabilities, and it’s been all about moving fast. I have already contributed to Llama4’s image…
Liked by Oleksandr Maksymets
-
Усі підбивають підсумки року, і я вирішила не залишатися осторонь. 2024 рік став для Fishka роком зростання, нових викликів та досягнень. Ось кілька…
Усі підбивають підсумки року, і я вирішила не залишатися осторонь. 2024 рік став для Fishka роком зростання, нових викликів та досягнень. Ось кілька…
Liked by Oleksandr Maksymets
-
Super excited to see this go live after months of work! Proud to be part of it!
Super excited to see this go live after months of work! Proud to be part of it!
Liked by Oleksandr Maksymets
Experience
Education
Licenses & Certifications
Volunteer Experience
-
-
Member of LC
AIESEC
- 1 year 10 months
Education
Complete 3 projects and worked in internship team for 2 months. Got leadership position in internship exchange team. Led team of 4 people to achievements in implementing outgoing process of interns. Team provided more than 50 interviews and signed 5 contracts during 2 weeks. Conducted National Leadership Development Seminar with 125 delegates and 25 business trainers. Communications with stakeholders and partners.
-
Publications
-
Integrating Egocentric Localization for More Realistic Point-Goal Navigation Agents
Recent work has presented embodied agents that can navigate to pointgoal targets in novel indoor environments with near-perfect accuracy. However,
these agents are equipped with idealized sensors for localization and take deterministic actions. This setting is practically sterile by comparison to the dirty reality of noisy sensors and actuations in the real world – wheels can slip, motion sensors have error, actuations can rebound. In this work, we take a step towards this
noisy reality…Recent work has presented embodied agents that can navigate to pointgoal targets in novel indoor environments with near-perfect accuracy. However,
these agents are equipped with idealized sensors for localization and take deterministic actions. This setting is practically sterile by comparison to the dirty reality of noisy sensors and actuations in the real world – wheels can slip, motion sensors have error, actuations can rebound. In this work, we take a step towards this
noisy reality, developing point-goal navigation agents that rely on visual estimates
of egomotion under noisy action dynamics. We find these agents outperform naive
adaptions of current point-goal agents to this setting as well as those incorporating
classic localization baselines. Further, our model conceptually divides learning
agent dynamics or odometry (where am I?) from task-specific navigation policy
(where do I want to go?). This enables a seamless adaption to changing dynamics (a different robot or floor type) by simply re-calibrating the visual odometry
model – circumventing the expense of re-training of the navigation policy. Our
agent was the runner-up in the PointNav track of CVPR 2020 Habitat Challenge. -
Habitat: A Platform for Embodied AI research (ICCV2019 Best Paper Award Nominees)
ICCV2019
We present Habitat, a new platform for research in embodied artificial intelligence (AI). Habitat enables training embodied agents (virtual robots) in highly efficient photorealistic 3D simulation, before transferring the learned skills to reality.
Specifically, Habitat consists of the following: 1. Habitat-Sim: a flexible, high-performance 3D simulator with configurable agents, multiple sensors, and generic 3D dataset handling (with built-in support for SUNCG, Matterport3D, Gibson…We present Habitat, a new platform for research in embodied artificial intelligence (AI). Habitat enables training embodied agents (virtual robots) in highly efficient photorealistic 3D simulation, before transferring the learned skills to reality.
Specifically, Habitat consists of the following: 1. Habitat-Sim: a flexible, high-performance 3D simulator with configurable agents, multiple sensors, and generic 3D dataset handling (with built-in support for SUNCG, Matterport3D, Gibson datasets). Habitat-Sim is fast--when rendering a scene from the Matterport3D dataset, Habitat-Sim achieves several thousand frames per second (fps) running single-threaded, and can reach over 10,000 fps multi-process on a single GPU, which is orders of magnitude faster than the closest simulator. 2. Habitat-API: a modular high-level library for end-to-end development of embodied AI algorithms--defining embodied AI tasks (eg navigation, instruction following, question answering), configuring and training embodied agents (via imitation or reinforcement learning, or via classic SLAM), and benchmarking using standard metrics. -
Embodied Question Answering in Photorealistic Environments with Point Cloud Perception (Oral)
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition
To help bridge the gap between internet vision-style problems and the goal of vision for embodied perception we instantiate a large-scale navigation task--Embodied Question Answering [1] in photo-realistic environments (Matterport 3D). We thoroughly study navigation policies that utilize 3D point clouds, RGB images, or their combination. Our analysis of these models reveals several key findings. We find that two seemingly naive navigation baselines, forward-only and random, are strong…
To help bridge the gap between internet vision-style problems and the goal of vision for embodied perception we instantiate a large-scale navigation task--Embodied Question Answering [1] in photo-realistic environments (Matterport 3D). We thoroughly study navigation policies that utilize 3D point clouds, RGB images, or their combination. Our analysis of these models reveals several key findings. We find that two seemingly naive navigation baselines, forward-only and random, are strong navigators and challenging to outperform, due to the specific choice of the evaluation setting presented by [1]. We find a novel loss-weighting scheme we call Inflection Weighting to be important when training recurrent models for navigation with behavior cloning and are able to out perform the baselines with this technique. We find that point clouds provide a richer signal than RGB images for learning obstacle avoidance, motivating the use (and continued study) of 3D deep learning models for embodied navigation.
-
Objectnav revisited: On evaluation of embodied agents navigating to objects
-
We revisit the problem of Object-Goal Navigation (ObjectNav). In its simplest form, ObjectNav is defined as the task of navigating to an object, specified by its label, in an
unexplored environment. In particular, the agent is initialized at a random location and pose in an environment and asked to find an instance of an object category, e.g. ‘find a chair’, by navigating to it.
Patents
-
Searching Online Social Networks Using Entity-based Embeddings
Filed US US20190114362A1
In one embodiment, a method includes receiving, from a client system associated with a user of an online social network, a search query for entities in the online social network, the search query containing one or more n-grams, generating a query embedding corresponding to the search query, where the query embedding represents the search query as a point in a d-dimensional embedding space, retrieving multiple entity embeddings corresponding to a plurality of entities, respectively, where each…
In one embodiment, a method includes receiving, from a client system associated with a user of an online social network, a search query for entities in the online social network, the search query containing one or more n-grams, generating a query embedding corresponding to the search query, where the query embedding represents the search query as a point in a d-dimensional embedding space, retrieving multiple entity embeddings corresponding to a plurality of entities, respectively, where each entity embedding represents the corresponding entity as a point in the d-dimensional embedding space, calculating, for each of the retrieved entity embeddings, a similarity metric between the query embedding and the entity embedding, ranking the entities based on their respective calculated similarity metrics, and sending, to the client system in response to the search query, instructions for presenting a search-results interface.
Other inventorsSee patent
Honors & Awards
-
ICCV2019 Best Paper Award Nominees
ICCV
-
Third Place Winner
National Ukrainian Olympiad in Informatics
-
Best Paper Award
International Scientific Conference TAAC in Kyiv(46 participants competed)
-
Opportunity Program Finalist
EducationUSA Network(80+ students competed)
Languages
-
English
Professional working proficiency
-
Ukrainian
Native or bilingual proficiency
Organizations
-
Association for Computing Machinery
Member of ACM
- Present
Recommendations received
2 people have recommended Oleksandr
Join now to viewMore activity by Oleksandr
-
Llama 4 is here! We have been working extremely hard to bring the latest version of llama models for the world to enjoy, both in open source, and…
Llama 4 is here! We have been working extremely hard to bring the latest version of llama models for the world to enjoy, both in open source, and…
Liked by Oleksandr Maksymets
-
On my way to DC with my team, to represent the Golden State at the 6th Ukraine Action Summit, the largest Ukrainian community advocacy event on the…
On my way to DC with my team, to represent the Golden State at the 6th Ukraine Action Summit, the largest Ukrainian community advocacy event on the…
Liked by Oleksandr Maksymets
Other similar profiles
Explore collaborative articles
We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.
Explore More