To overcome challenges posed by escalating environmental pollution and climate change, the combined economic and emission dispatch problem is proposed to balance economic efficiency with emission cost. The primary objective of the problem is to ensure that emissions are minimized while optimal economic costs are achieved simultaneously. However, due to the nonlinear and nonconvex characteristics of the model, the optimization is confronted with many difficulties. Hence, an innovative improved reinforcement learning-based differential evolution algorithm is proposed in this article, with reinforcement learning seamlessly integrated into the differential evolution algorithm. Q-learning from reinforcement learning technique is utilized to dynamically adjust parameter settings and select appropriate mutation strategies, thereby boosting the algorithm's adaptability and overall performance. The effectiveness of the proposed algorithm is tested on thirty testing functions and combined economic and emission dispatch problems in comparison with the other five algorithms. According to the experimental results of testing functions, superior performance is consistently achieved by the proposed algorithm, with the highest adaptability exhibited and an average ranking of 1.4167. Its superiority is further demonstrated through Wilcoxon tests on results of testing functions and combined economic and emission dispatch problems with the proportion of 100%, and the proposed algorithm is significantly better than other algorithms at a 0.05 significance level. The superiority of the proposed algorithm in optimizing combined economic and emission dispatch problems demonstrates that the proposed algorithm is shown to be adaptable to complex optimization environments, which proves useful for industrial applications and artificial intelligence.