{"id":4551,"date":"2025-10-31T10:12:43","date_gmt":"2025-10-31T14:12:43","guid":{"rendered":"https:\/\/oge.mit.edu\/msrp\/?post_type=profiles&#038;p=4551"},"modified":"2025-12-09T12:05:01","modified_gmt":"2025-12-09T17:05:01","slug":"nandan-sarkar-2","status":"publish","type":"profiles","link":"https:\/\/oge.mit.edu\/msrp\/profiles\/nandan-sarkar-2\/","title":{"rendered":"Nandan Sarkar"},"content":{"rendered":"<div class=\"wp-block-image\">\n<figure class=\"alignleft size-full is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"2560\" height=\"2560\" src=\"https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-scaled.jpg\" alt=\"\" class=\"wp-image-4552\" style=\"width:200px;height:auto\" srcset=\"https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-scaled.jpg 2560w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-300x300.jpg 300w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-1024x1024.jpg 1024w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-150x150.jpg 150w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-768x768.jpg 768w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-1536x1536.jpg 1536w, https:\/\/oge.mit.edu\/msrp\/wp-content\/uploads\/sites\/2\/2025\/11\/SarkarNadan-edited-2048x2048.jpg 2048w\" sizes=\"auto, (max-width: 2560px) 100vw, 2560px\" \/><\/figure>\n<\/div>\n\n\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<p><strong>MIT Department:<\/strong> Electrical Engineering and Computer Science<br><strong>Faculty Mentor<\/strong>: Prof. Yoon Kim<br><strong>Research Supervisor:<\/strong> Abbas Zeitoun<br><strong>Undergraduate Institution:<\/strong> Yale University<br><strong>Website<\/strong>:<\/p>\n<\/div><\/div>\n\n\n\n<div style=\"height:0px\" aria-hidden=\"true\" class=\"wp-block-spacer\"><\/div>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Biography<\/strong><\/h4>\n\n\n\n<p>Nandan Sarkar is a senior from Nashville, Tennessee, studying Computer Science andApplied Math at Yale. He is very interested in machine learning and is committed to pursuing a Ph.D. in this field. At Yale, he is an undergraduate researcher in the YaleNLP lab underProfessor Arman Cohan. For the past year\u2014and this summer at MIT\u2014he has been a member of the Computation and Language Lab under Professor Yoon Kim. Previously, Nandan conducted research at Vanderbilt on human-computer interaction and augmented reality. Since freshman year, Nandan has been deeply involved in Code Haven, a student-run organization dedicated to expanding computer science education for middle school students in New Haven. Outside of school, Nandan enjoys playing squash, watching detective shows, and playing poker with friends. He is also a big soccer and football fan (a die-hard fan of Barcelona and the Tennessee Titans) and loves country music and hot chicken.<\/p>\n\n\n\n<h4 class=\"wp-block-heading\"><strong>Abstract<\/strong><\/h4>\n\n\n\n<p class=\"has-text-align-center\"><strong>Learning to Reason from Inferred Steps: An EM Framework for Enhancing LL Reasoning<\/strong><\/p>\n\n\n\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<p class=\"has-text-align-center\"><strong>Nandan Sarkar<sup>1<\/sup>, Abbas Zeitoun<sup>2<\/sup>, and Yoon Kim<sup>2<\/sup><\/strong><\/p>\n\n\n\n<div class=\"wp-block-group\"><div class=\"wp-block-group__inner-container is-layout-constrained wp-block-group-is-layout-constrained\">\n<div class=\"wp-block-group is-vertical is-content-justification-center is-layout-flex wp-container-core-group-is-layout-4b2eccd6 wp-block-group-is-layout-flex\">\n<p class=\"has-text-align-center\"><sup>1<\/sup>Department of Computer Science, Yale University<\/p>\n\n\n\n<p><sup>2<\/sup>Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology<\/p>\n\n\n\n<p class=\"has-text-align-center\"><\/p>\n<\/div>\n<\/div><\/div>\n<\/div><\/div>\n<\/div><\/div>\n<\/div><\/div>\n<\/div><\/div>\n\n\n\n<p>Large language models (LLMs) have shown impressive capabilities in solving complex tasks that require multi-step reasoning, such as those in mathematics, science, and logic. However, training models to produce high-quality reasoning traces often involves supervised fine-tuning on teacher-generated step-by-step explanations or reinforcement learning using verifiable rewards. In this work, we aim to enhance reasoning capabilities, particularly on tasks that lack verifiable rewards, by enabling the model to better interpret and internalize reasoning traces. We introduce an Expectation-Maximization (EM) framework that treats intermediate reasoning steps as latent variables. We begin by fine-tuning OpenThinker-7B, a model exposed to diverse reasoning data, to predict missing steps in reasoning traces. Then, during the E-step, we sample latent intermediate steps for partially masked traces from the model\u2019s predictions. In the M-step, the model is fine-tuned on these completed traces, using its own inferred reasoning to iteratively improve. This self-training approach enables the model to enhance its reasoning abilities without heavily relying on external supervision. We evaluate our method on a range of challenging reasoning benchmarks, including AMC and AIME math competitions, GPQA science questions, and other logic-intensive tasks. Preliminary results suggest that our EM-style framework improves both reasoning coherence and final answer accuracy, particularly under constrained data budgets.<\/p>\n","protected":false},"featured_media":4552,"template":"","profile_category":[23],"class_list":["post-4551","profiles","type-profiles","status-publish","has-post-thumbnail","hentry","profile_category-2025-interns"],"acf":[],"_links":{"self":[{"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/profiles\/4551","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/profiles"}],"about":[{"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/types\/profiles"}],"version-history":[{"count":2,"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/profiles\/4551\/revisions"}],"predecessor-version":[{"id":4848,"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/profiles\/4551\/revisions\/4848"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/media\/4552"}],"wp:attachment":[{"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/media?parent=4551"}],"wp:term":[{"taxonomy":"profile_category","embeddable":true,"href":"https:\/\/oge.mit.edu\/msrp\/wp-json\/wp\/v2\/profile_category?post=4551"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}