My name is Satyajit Bhowmick(fondly known as 'Jit'). I'm a computer science graduate, an avid programmer and a tech buff. My research interest lies in the area of data science and machine learning. I believe in 'learning things by doing' and by doing so I strive for excellence in my area of interest. Most importantly I try to improve myself everyday to be a better version of me. Find more details about me from this page. Keep visit to this page as I update it very often.
University of Cincinnati, USA Thesis: A Fog-based Cloud Paradigm for Time Sensitive Applications Specialization: Data Science, Cloud Computing, Big data, Machine Learning.
West Bengal University of Technology, India Final Project: Estimation of Hand Force from Surface Electromyography Signals using Artificial Neural Network
* Maintain and improve the SuperComm2 application system and webpage for high-volume big data repository of SuperComm2 Command Center. * Write robust applications to convert high frequency raw machine data(.SC2) to other human readable formats like GH5, THD, TDMS, XML. * Created data pipeline to transfer real-time high frequency data to AWS S3 buckets while maintaining the sanity of data. * Write multithreaded python scripts that run across multiple servers. * Develop and maintain proprietary python data analysis framework called PyDAX. * Responsible for analyzing various cross-functional, multi-platform applications systems enforcing Python best practices and provide guidance in making long term architectural design decisions. * Used Spark to perform data exploration and mining on real Apache web server log files. * Consumed consumer geo data from Caterpillar feedback stream and rendered locations on map. * Efficiently used clustering algorithms like SVM, KNN. * Successfully used python Scikit-learn for predictive modelling. * Used Hadoop and MapReduce for Caterpillar equipment channel data analysis. * Developed a Restful service that mines more than 70TB of data and responds to user requests via APIS using Python Flask framework, Celery, RabbitMQ, Couchbase, Tornado and Nginx. * Built web scraper and extracted data from web pages using python and BeautifulSoup. * Use Python modules such as requests, urllib, urllib2 for web crawling and JavaScript as well. * Successfully set up an intranet mail server using hMailServer and built email crawler using python and eml_parser to parse email body. * Used python’s SMTP protocol to send customized email. * Create RESTFUL API's for several of our Intranet applications using open source software packages. * Develop remote integration with third party platforms by using RESTful web services. * Use Python to extract information from XML files. * Actively involved in the initial software development life cycle (SDLC) of requirement gathering and in suggesting system configuration specifications during client interaction. * Generate various graphs for business decision making using Python matplotlib library. * Expertise in data visualization with Tableau. Embed tableau dashboards on SuperComm2 Website to ensure better and dynamic user experience. * Design and create the database tables and write SQL queries to access Oracle. * Use Couchbase and Redis to reduce overhead and respond quickly to API requests. * Use AWS(Amazon Web services) EMR for improved efficiency of storage. * Utilize Python Libraries like Boto3, NumPy, Pandas for AWS. * Use Celery as task queue and RabbitMQ as messaging broker to execute asynchronous tasks. * Comfortable in creating and maintaining Docker containers. * Involved in writing Shell scripts to automate tasks and application specific syncs / backups and other schedulers. * Design, model and optimize relational databases tables in MySQL. * Perform requirements gathering and work closely with the architect in designing. * Use agile development tools and methodologies in Python, Git, PyCharm, code review. * Work on data extraction, data mapping and data insertion; the process of data migration. * Hands on experience in working with the Cassandra Query Language (CQL) for querying the data present in the Cassandra. * Resolved several hidden bugs caused by complicated multithreading issues such as race conditions caused by asynchronous events. * Designed user interactive web pages as the front-end part of the web application using various web technologies like HTML, JavaScript, jQuery, AJAX and implemented CSS for better appearance and feel, * Deployed projects into Heroku and used GitHub version control system.
* Analyzed datasets more than 1TB in size to find trends. * Used Hadoop and python for twitter trend analysis. * Created fog computing architecture and programming model for efficient analysis of time sensitive data. * Used machine learning algorithms like SVM, KNN, Decision Tree, K-Means for intelligent data analysis. * Used Artificial Neural Network for predictive modelling.