Rishab Sharma

Rishab Sharma

Mumbai, Maharashtra, India
7K followers 500+ connections

About

Visit https://personate.ai



I am a self-taught data scientist and visual…

Services

Activity

Join now to see all activity

Experience

  • Personate.ai Graphic
  • -

    Mumbai, Maharashtra, India

  • -

    Mumbai, Maharashtra, India

  • -

  • -

    Ahmedabad, Gujarat, India

  • -

    San Francisco Bay Area

  • -

    Banglore

  • -

    Dehradun, Uttarakhand, India

Education

Licenses & Certifications

Volunteer Experience

  • Marathon Runner

    ONGC

    Ran for the cause of Tree Plantation in our city of Dehra Dun

Publications

  • Salient Image Matting

    Cornell

    In this paper, we propose an image matting framework called Salient Image Matting to estimate the per-pixel opacity value of the most salient foreground in an image. To deal with a large amount of semantic diversity in images, a trimap is conventionally required as it provides important guidance about object semantics to the matting process. However, creating a good trimap is often expensive and timeconsuming. The SIM framework simultaneously deals with the challenge of learning a wide range of…

    In this paper, we propose an image matting framework called Salient Image Matting to estimate the per-pixel opacity value of the most salient foreground in an image. To deal with a large amount of semantic diversity in images, a trimap is conventionally required as it provides important guidance about object semantics to the matting process. However, creating a good trimap is often expensive and timeconsuming. The SIM framework simultaneously deals with the challenge of learning a wide range of semantics and salient object types in a fully automatic and an end to end manner. Specifically, our framework is able to produce accurate alpha mattes for a wide range of foreground objects and cases where the foreground class, such as human, appears in a very different context than the train data directly from an RGB input. This is done by employing a salient object detection model to produce a trimap of the most salient object in the image in order to guide the matting model about higher-level object semantics. Our framework leverages large amounts of coarse annotations coupled with a heuristic trimap generation scheme to train the trimap prediction network so it can produce trimaps for arbitrary foregrounds. Moreover, we introduce a multi-scale fusion architecture for the task of matting to better capture finer, low-level opacity semantics. With high-level guidance provided by the trimap network, our framework requires only a fraction of expensive matting data as compared to other automatic methods while being able to produce alpha mattes for a diverse range of inputs. We demonstrate our framework on a range of diverse images and experimental results show our framework compares favourably against state of art matting methods without the need for a trimap

    Other authors
    See publication
  • Employing Differentiable Neural Computers for Image Captioning and Neural Machine Translation

    Elsevier - Procedia Computer Science Volume 173, 2020, Pages 234-244

    In the history of artificial neural networks, LSTMs have proved to be a high-performance architecture at sequential data learning. Although LSTMs are remarkable in learning sequential data but are limited in their ability to learn long-term dependencies and representation of certain data structures because of the lack of external memory. In this paper, we tackled two main tasks, one is language translation and other is image captioning. We approached the problem of language translation by…

    In the history of artificial neural networks, LSTMs have proved to be a high-performance architecture at sequential data learning. Although LSTMs are remarkable in learning sequential data but are limited in their ability to learn long-term dependencies and representation of certain data structures because of the lack of external memory. In this paper, we tackled two main tasks, one is language translation and other is image captioning. We approached the problem of language translation by leveraging the capabilities of the recently developed DNC architectures. Here we modified the DNC architecture by including dual neural controllers instead of one and an external memory module. Inside our controller, we employed a neural network with memory-augmentation which differs from the original differentiable neural computer, we implemented a dual controller’s system in which one controller is for encoding the query sequence whereas another controller is for decoding the translated sequences. During the encoding cycle, new inputs are read and the memory is updated accordingly. In the decoding cycle, the memory is protected from any writing from the decoding controller. Thus, the decoder phase generates a translated sequence at a time step. Therefore, the proposed dual controller neural network with memory-augmentation is then trained and tested on the Europarl dataset. For the image captioning task, our architecture is inspired by an end-to-end image captioning model where CNN’s output is passed to RNN as input only once and the RNN generates words depending on the input. We trained our DNC captioning model on 2015 MSCOCO dataset. In the end, we compared and shows the superiority of our architecture as compared to conventionally used LSTM and NTM architectures.

    Other authors
    See publication
  • AlphaNet: An Attention Guided Deep Network for Automatic Image Matting

    IEEE COINS 2020

    In this paper, we propose an end to end solution for image matting i.e high-precision extraction of foreground objects from natural images. Image matting and background detection can be achieved easily through chroma keying in a studio setting when the background is either pure green or blue. Nonetheless, image matting in natural scenes with complex and uneven depth backgrounds remains a tedious task that requires human intervention. To achieve complete automatic foreground extraction in…

    In this paper, we propose an end to end solution for image matting i.e high-precision extraction of foreground objects from natural images. Image matting and background detection can be achieved easily through chroma keying in a studio setting when the background is either pure green or blue. Nonetheless, image matting in natural scenes with complex and uneven depth backgrounds remains a tedious task that requires human intervention. To achieve complete automatic foreground extraction in natural scenes, we propose a method that assimilates semantic segmentation and deep image matting processes into a single network to generate detailed semantic mattes for image composition task. The contribution of our proposed method is two-fold, firstly it can be interpreted as a fully automated semantic image matting method and secondly as a refinement of existing semantic segmentation models. We propose a novel model architecture as a combination of segmentation and matting that unifies the function of upsampling and downsampling operators with the notion of attention. As shown in our work, attention guided downsampling and upsampling can extract high-quality boundary details, unlike other normal downsampling and upsampling techniques. For achieving the same, we utilized an attention guided encoder-decoder framework which does unsupervised learning for generating an attention map adaptively from the data to serve and direct the upsampling and downsampling operators. We also construct a fashion e-commerce focused dataset with high-quality alpha mattes to facilitate the training and evaluation for image matting.

    Other authors
    See publication
  • Retrieving Similar E-Commerce Images Using Deep Learning

    arXiv.org

    In this paper, we propose a deep convolutional neural network for learning the embeddings of images in order to capture the notion of visual similarity. We present a deep siamese architecture that when trained on positive and negative pairs of images learn an embedding that accurately approximates the ranking of images in order of visual similarity notion. We also implement a novel loss calculation method using an angular loss metrics based on the problems requirement. The final embedding of…

    In this paper, we propose a deep convolutional neural network for learning the embeddings of images in order to capture the notion of visual similarity. We present a deep siamese architecture that when trained on positive and negative pairs of images learn an embedding that accurately approximates the ranking of images in order of visual similarity notion. We also implement a novel loss calculation method using an angular loss metrics based on the problems requirement. The final embedding of the image is combined representation of the lower and top-level embeddings. We used fractional distance matrix to calculate the distance between the learned embeddings in n-dimensional space. In the end, we compare our architecture with other existing deep architecture and go on to demonstrate the superiority of our solution in terms of image retrieval by testing the architecture on four datasets. We also show how our suggested network is better than the other traditional deep CNNs used for capturing fine-grained image similarities by learning an optimum embedding.

    Other authors
    See publication
  • Neural network for Image Classification

    International Journal of Innovative Computer Science and Engineerinfg

    As observed machine learning, computer vision techniques and other computer science algorithms cannot compete the human level of intelligence in pattern recognition such as hand written digits and traffic signs. But here we have reviewed a biologically plausible deep neural network architecture which can make it possible using a fully parameterizable GPU implementation deep neural network independent of the pre-wired feature extractors designing, which are rather learned in a supervised way. In…

    As observed machine learning, computer vision techniques and other computer science algorithms cannot compete the human level of intelligence in pattern recognition such as hand written digits and traffic signs. But here we have reviewed a biologically plausible deep neural network architecture which can make it possible using a fully parameterizable GPU implementation deep neural network independent of the pre-wired feature extractors designing, which are rather learned in a supervised way. In this method tiny fields of winner neurons gives sparsely connected neural layers which leads to huge network depth as found in human like species between retina and visual cortex. The winning neurons are trained on many columns of deep neurons to attain expertise on pre-processed inputs in many different ways after which their predictions are averaged. Also GPU used, enables the models to be trained faster than usual. Upon testing the proposed method over MNIST handwriting data it achieves a near-human performance. Upon considering traffic sign recognition, our architecture has an upper hand by a factor of two. We also tried to improve the state-of-theart on a huge amount of common image classification benchmarks.

    Other authors
    See publication

Courses

  • HTML and CSS

    -

  • Julia

    -

  • Machine Learning

    -

  • Make a website

    -

  • OpenCV

    -

  • Python

    -

  • Python-Django

    -

  • Python-Flask

    -

  • Robotics and IOT

    -

Projects

  • snack search

    A machine Learning app with a multiple Algorithms Engine Support
    One of the main uses of computers is to help us solve problems quickly and effectively. And a problem we often run into is figuring out what to eat, or what to make. This problem is solvable using data and recommendation engines.

    Recommendation engines work on two levels. The first level is on the personal level. Let's say you create a dataset of foods and rank how much you enjoy or dislike them, 1-10. Given an unseen…

    A machine Learning app with a multiple Algorithms Engine Support
    One of the main uses of computers is to help us solve problems quickly and effectively. And a problem we often run into is figuring out what to eat, or what to make. This problem is solvable using data and recommendation engines.

    Recommendation engines work on two levels. The first level is on the personal level. Let's say you create a dataset of foods and rank how much you enjoy or dislike them, 1-10. Given an unseen food and its set of features (such as the inclusion of ingredients, or perhaps the percentage of that meal the ingredient takes up). A machine learning algorithm figure out if and how much you'd like it. The other way recommendation engines work is on the group level. A machine learning algorithm should be able to recommend new foods to you, given a set of people who share your similar food preferences.

    The goal for this project is to build a system that allows you to identify and then recommend, recipes you're likely to enjoy.

    I have used Flask Microframework to serve my app.

    See project
  • Neural Machine Translation Chatbot

    - Present

    Developed a Chatbot based on Neural Machine Translation architecture of GNMT ( Google Translate ) to chat about Stackoverflow data and other programming question and answer discussion Forums.

  • OCR on Android

    An android app to recognise a hand Written Digit using a Flask , Tensorflow Backend , Model Designed using The coolest and simplest Keras

    See project
  • All in One Emergency Services Portal

    created an all in one web portal for any kind of emergency services including medical , 🔥 fire , 👮 police , water , internet attack like DDOS and SQL injection , etc. along with ambulance tracking , GPS location sharing and smart help mode.

    See project
  • Line following robo car and A Bluetooth and Wifi Controlled Robo Car

    Under this project we designed a Line following robo car along with a Bluetooth controlled and a wifi controlled Robo Car

    Other creators
    See project
  • Cross a Crator

    India has sent its rover for an expedition to the red planet Mars. The mission is to collect samples to determine if life sustaining factors exist in this planet. While returning to the Base Station from a different route the rover has encountered a huge crater that it needs to cross. The crater has two paths comprising of cavities along the way. These cavities need to be filled using appropriate boulders by taking a feed from the nearest satellite to make them traversable.

    The above…

    India has sent its rover for an expedition to the red planet Mars. The mission is to collect samples to determine if life sustaining factors exist in this planet. While returning to the Base Station from a different route the rover has encountered a huge crater that it needs to cross. The crater has two paths comprising of cavities along the way. These cavities need to be filled using appropriate boulders by taking a feed from the nearest satellite to make them traversable.

    The above scenario has been simplified and abstracted as an arena for this theme. The arena represents a crater and comprises of two partially traversable bridges, Bridge 1 and Bridge 2 with cavities at random positions leading to the Base Station. The rover takes the feed from a camera directly above it that guides it towards filling the cavities using conical structures and navigating the bridge. Navigating each of the bridges involve different challenges. The rover has to traverse using one of these bridges and reach the Base Station.

    Other creators
    See project
  • pratyeti

    -

    प्रत्यति(Believe)
    App Summary
    Our app aims at adding another feature to the current navigation system used by people in day-to- day life. The feature, as we call it is safe-navigation. It allows you to not only find the shortest route from an origin to destination but also the safest- based on the trends in crimes recorded along all possible routes over a certain period of time. For this, we plan to create clusters based upon latitudes and longitudes of recorded crimes. We will assign…

    प्रत्यति(Believe)
    App Summary
    Our app aims at adding another feature to the current navigation system used by people in day-to- day life. The feature, as we call it is safe-navigation. It allows you to not only find the shortest route from an origin to destination but also the safest- based on the trends in crimes recorded along all possible routes over a certain period of time. For this, we plan to create clusters based upon latitudes and longitudes of recorded crimes. We will assign these clusters a danger index based upon the types, density and date of these recorded crimes. Then we will use these clusters to calculate a heuristic function for an AI-algorithm to traverse each possible step of all possible routes and calculate a final safety-index of each route. These safety-indices will be assigned to the routes on-demand and hence will change with the time at which a user plans to use the app, as a route might be comparatively less safe at certain times than others.
    But very few, if any, recorded FIRs are ever made public. Also, due of lack of transparency in the way crime is reported to the police, these FIRs are easy to manipulate and hence any public records of these generated by the current system is lacking in reliability. Thus to solve this problem, we plan to create a blockchain architecture of recording these FIRs so that a public ledger of every FIR recorded is present on a server at every police station which is a part of our blockchain network. This will ensure that no individual-be it a person with authority or a common man-can manipulate this information. Using this information we plan on generating a reliable database that can be used by our cluster of AI-models to give a new dimension to current standards of navigation.
    If successfully implemented, we can easily migrate the current records of FIRs to our blockchain for preserving their authenticity while providing a method to record all the new FIRs to the blockchain network directly.

    Other creators
    • harshit singhal
    • pranav singhal
    • arvind Kalra
    See project
  • Disease Recognition and Epidemic Detector

    -

    In this project we do disease recognition through two possible and feasible ways of Image Processing and symptoms analysis using state of the art Convolutional Neural Network , KNNeighbours and Support Vector Classifier of Deep learning and machine learning techniques . We aim to cluster patients with similar symptoms located closely on geo-location using ML clustering algorithms and plot the infected polygons on a map template for analysis and quick actions. Further it could be extended to…

    In this project we do disease recognition through two possible and feasible ways of Image Processing and symptoms analysis using state of the art Convolutional Neural Network , KNNeighbours and Support Vector Classifier of Deep learning and machine learning techniques . We aim to cluster patients with similar symptoms located closely on geo-location using ML clustering algorithms and plot the infected polygons on a map template for analysis and quick actions. Further it could be extended to identification of blood donation requirement areas and also medicine prespcription based on the data and provide relevant surveys.

    See project
  • Chatbot using Tensorflow

    -

    Try it at
    https://rishabbot.herokuapp.com/

    See project
  • Webcam Object Detection using Tensorflow

    -

    A tensorflow model to detect objects in a webcam Video using OpenCV capable of localizing and identifying multiple objects in a single image

    See project
  • Geospatial database map plotting

    -

    MAP PLOTTING USING GEOSPATIAL DATABASE is 2 way project in which a geospatial database can be plotted on a map depicting the various polygons which refer to the lease area of a company along with the point data depicting the wells and boreholes.The geospatial data consists of shape files (contains the array of lat-long points of lease polygon) and the well lat-long in degree.The map plotting is done using Interactive python and its libraries ( matplotlib , basemap , _proj4(for degree to…

    MAP PLOTTING USING GEOSPATIAL DATABASE is 2 way project in which a geospatial database can be plotted on a map depicting the various polygons which refer to the lease area of a company along with the point data depicting the wells and boreholes.The geospatial data consists of shape files (contains the array of lat-long points of lease polygon) and the well lat-long in degree.The map plotting is done using Interactive python and its libraries ( matplotlib , basemap , _proj4(for degree to cartesian conversion) , pyplot , mpl_toolkits). The other way is collecting geospatial data by manual map marking. In this method we use google map api in javascript and connect it to a PostgreSQL using SQLAlchemy and a Python microframework of Jinja2 (Flask).The “OnClick” events on the map surface are handled by the JS code and responses are stored in the PostgreSQL database , thus enabling the creation of a geospatial database with lat-long points which refer to the earlier mentioned parameter of boreholes attributes.Thus we have developed a 2-way approach to interact with a geospatial database enabling both its creation and its visualisation.

    Other creators
    • akanshi rawat
    • Mohammad zohaib
    See project

Languages

  • English

    -

  • Hindi

    -

  • Sanskrit

    -

More activity by Rishab

View Rishab’s full profile

  • See who you know in common
  • Get introduced
  • Contact Rishab directly
Join to view full profile

Other similar profiles

Explore collaborative articles

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Explore More

Add new skills with these courses