Oleksandr Maksymets

Oleksandr Maksymets

San Francisco Bay Area
3K followers 500+ connections

About

Tech Lead, Research Engineer at Facebook AI Research (FAIR) with 10+ years of industry…

Articles by Oleksandr

Activity

Join now to see all activity

Experience

  • Facebook AI Graphic

    Facebook AI

    Menlo Park

  • -

    Menlo Park, California, United States

  • -

    London, United Kingdom

  • -

  • -

  • -

  • -

  • -

Education

Licenses & Certifications

Volunteer Experience

  • ACM, Association for Computing Machinery Graphic

    Member

    ACM, Association for Computing Machinery

    - 2 years 1 month

  • AIESEC Graphic

    Member of LC

    AIESEC

    - 1 year 10 months

    Education

    Complete 3 projects and worked in internship team for 2 months. Got leadership position in internship exchange team. Led team of 4 people to achievements in implementing outgoing process of interns. Team provided more than 50 interviews and signed 5 contracts during 2 weeks. Conducted National Leadership Development Seminar with 125 delegates and 25 business trainers. Communications with stakeholders and partners.

  • AEGEE / European Students'​ Forum Graphic

    Member

    AEGEE / European Students'​ Forum

    - 2 years

    Arts and Culture

Publications

  • Integrating Egocentric Localization for More Realistic Point-Goal Navigation Agents

    Recent work has presented embodied agents that can navigate to pointgoal targets in novel indoor environments with near-perfect accuracy. However,
    these agents are equipped with idealized sensors for localization and take deterministic actions. This setting is practically sterile by comparison to the dirty reality of noisy sensors and actuations in the real world – wheels can slip, motion sensors have error, actuations can rebound. In this work, we take a step towards this
    noisy reality…

    Recent work has presented embodied agents that can navigate to pointgoal targets in novel indoor environments with near-perfect accuracy. However,
    these agents are equipped with idealized sensors for localization and take deterministic actions. This setting is practically sterile by comparison to the dirty reality of noisy sensors and actuations in the real world – wheels can slip, motion sensors have error, actuations can rebound. In this work, we take a step towards this
    noisy reality, developing point-goal navigation agents that rely on visual estimates
    of egomotion under noisy action dynamics. We find these agents outperform naive
    adaptions of current point-goal agents to this setting as well as those incorporating
    classic localization baselines. Further, our model conceptually divides learning
    agent dynamics or odometry (where am I?) from task-specific navigation policy
    (where do I want to go?). This enables a seamless adaption to changing dynamics (a different robot or floor type) by simply re-calibrating the visual odometry
    model – circumventing the expense of re-training of the navigation policy. Our
    agent was the runner-up in the PointNav track of CVPR 2020 Habitat Challenge.

    See publication
  • Habitat: A Platform for Embodied AI research (ICCV2019 Best Paper Award Nominees)

    ICCV2019

    We present Habitat, a new platform for research in embodied artificial intelligence (AI). Habitat enables training embodied agents (virtual robots) in highly efficient photorealistic 3D simulation, before transferring the learned skills to reality.
    Specifically, Habitat consists of the following: 1. Habitat-Sim: a flexible, high-performance 3D simulator with configurable agents, multiple sensors, and generic 3D dataset handling (with built-in support for SUNCG, Matterport3D, Gibson…

    We present Habitat, a new platform for research in embodied artificial intelligence (AI). Habitat enables training embodied agents (virtual robots) in highly efficient photorealistic 3D simulation, before transferring the learned skills to reality.
    Specifically, Habitat consists of the following: 1. Habitat-Sim: a flexible, high-performance 3D simulator with configurable agents, multiple sensors, and generic 3D dataset handling (with built-in support for SUNCG, Matterport3D, Gibson datasets). Habitat-Sim is fast--when rendering a scene from the Matterport3D dataset, Habitat-Sim achieves several thousand frames per second (fps) running single-threaded, and can reach over 10,000 fps multi-process on a single GPU, which is orders of magnitude faster than the closest simulator. 2. Habitat-API: a modular high-level library for end-to-end development of embodied AI algorithms--defining embodied AI tasks (eg navigation, instruction following, question answering), configuring and training embodied agents (via imitation or reinforcement learning, or via classic SLAM), and benchmarking using standard metrics.

    See publication
  • Embodied Question Answering in Photorealistic Environments with Point Cloud Perception (Oral)

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition

    To help bridge the gap between internet vision-style problems and the goal of vision for embodied perception we instantiate a large-scale navigation task--Embodied Question Answering [1] in photo-realistic environments (Matterport 3D). We thoroughly study navigation policies that utilize 3D point clouds, RGB images, or their combination. Our analysis of these models reveals several key findings. We find that two seemingly naive navigation baselines, forward-only and random, are strong…

    To help bridge the gap between internet vision-style problems and the goal of vision for embodied perception we instantiate a large-scale navigation task--Embodied Question Answering [1] in photo-realistic environments (Matterport 3D). We thoroughly study navigation policies that utilize 3D point clouds, RGB images, or their combination. Our analysis of these models reveals several key findings. We find that two seemingly naive navigation baselines, forward-only and random, are strong navigators and challenging to outperform, due to the specific choice of the evaluation setting presented by [1]. We find a novel loss-weighting scheme we call Inflection Weighting to be important when training recurrent models for navigation with behavior cloning and are able to out perform the baselines with this technique. We find that point clouds provide a richer signal than RGB images for learning obstacle avoidance, motivating the use (and continued study) of 3D deep learning models for embodied navigation.

    See publication
  • Objectnav revisited: On evaluation of embodied agents navigating to objects

    -

    We revisit the problem of Object-Goal Navigation (ObjectNav). In its simplest form, ObjectNav is defined as the task of navigating to an object, specified by its label, in an
    unexplored environment. In particular, the agent is initialized at a random location and pose in an environment and asked to find an instance of an object category, e.g. ‘find a chair’, by navigating to it.

    See publication

Patents

  • Searching Online Social Networks Using Entity-based Embeddings

    Filed US US20190114362A1

    In one embodiment, a method includes receiving, from a client system associated with a user of an online social network, a search query for entities in the online social network, the search query containing one or more n-grams, generating a query embedding corresponding to the search query, where the query embedding represents the search query as a point in a d-dimensional embedding space, retrieving multiple entity embeddings corresponding to a plurality of entities, respectively, where each…

    In one embodiment, a method includes receiving, from a client system associated with a user of an online social network, a search query for entities in the online social network, the search query containing one or more n-grams, generating a query embedding corresponding to the search query, where the query embedding represents the search query as a point in a d-dimensional embedding space, retrieving multiple entity embeddings corresponding to a plurality of entities, respectively, where each entity embedding represents the corresponding entity as a point in the d-dimensional embedding space, calculating, for each of the retrieved entity embeddings, a similarity metric between the query embedding and the entity embedding, ranking the entities based on their respective calculated similarity metrics, and sending, to the client system in response to the search query, instructions for presenting a search-results interface.

    Other inventors
    See patent

Honors & Awards

  • ICCV2019 Best Paper Award Nominees

    ICCV

  • Third Place Winner

    National Ukrainian Olympiad in Informatics

  • Best Paper Award

    International Scientific Conference TAAC in Kyiv(46 participants competed)

  • Opportunity Program Finalist

    EducationUSA Network(80+ students competed)

Languages

  • English

    Professional working proficiency

  • Ukrainian

    Native or bilingual proficiency

Organizations

  • Association for Computing Machinery

    Member of ACM

    - Present

Recommendations received

More activity by Oleksandr

View Oleksandr’s full profile

  • See who you know in common
  • Get introduced
  • Contact Oleksandr directly
Join to view full profile

Other similar profiles

Explore collaborative articles

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Explore More

Add new skills with these courses