没有合适的资源?快使用搜索试试~ 我知道了~
基于Q学习算法的集装箱堆场翻箱落位优选1
需积分: 0 0 下载量 188 浏览量
2022-08-03
23:14:07
上传
评论
收藏 2.39MB PDF 举报
温馨提示
试读
49页
摘要港口是综合交通运输网络中最重要的水陆交通枢纽,是物流供应链中最大的集货中心,随着货物集装箱化的加深,码头集装箱吞吐量逐渐增多,集装箱堆场的堆存压力日益加重,
资源详情
资源评论
资源推荐
专 业 学 位 硕 士 学 位 论 文
基于 Q 学习算法的集装箱堆场翻箱落位优选
Reshuffle Optimization in Container Terminal based on
Q Learning Algorithm
作 者 姓 名: 叶 倩 倩
工 程 领 域: 水利工程
学 号: 31706200
指 导 教 师: 周 鹏 飞
完 成 日 期: 2019 年 6 月
大连理工大学
Dalian University of Technology
万方数据
大连理工大学学位论文独创性声明
作者郑重声明:所呈交的学位论文,是本人在导师的指导下进行研究
工作所取得的成果。尽我所知,除文中已经注明引用内容和致谢的地方外,
本论文不包含其他个人或集体已经发表的研究成果,也不包含其他已申请
学位或其他用途使用过的成果。与我一同工作的同志对本研究所做的贡献
均已在论文中做了明确的说明并表示了谢意。
若有不实之处,本人愿意承担相关法律责任。
学位论文题目:
作 者 签 名 : 日期: 年 月 日
万方数据
大连理工大学专业学位硕士学位论文
- I -
摘 要
港口是综合交通运输网络中最重要的水陆交通枢纽,是物流供应链中最大的集货中
心,随着货物集装箱化的加深,码头集装箱吞吐量逐渐增多,集装箱堆场的堆存压力日
益加重,堆场资源变得紧张。为提高客户服务质量,提升港口的综合竞争力,优化集装
箱码头内的调度管理至关重要。而堆场翻箱率是集装箱码头调度管理的重要指标之一,
提箱过程中的翻箱落位优化可有效降低二次及二次以上翻箱率。
本文针对集装箱堆场的翻箱问题,首先从进口箱翻箱、出口箱翻箱和移箱时翻箱三
种情况对堆场翻箱问题进行分类分析,然后从内部因素、外部因素与其他不可抗因素等
方向分析翻箱产生原因,提出从堆场管理方面降低堆场翻箱量的方法,最后确定本文的
研究对象为进口箱,研究进口箱的翻箱落位优选以降低堆场翻箱率。
堆场贝位翻箱问题以最小化贝位的二次及二次以上翻箱率为优化目标,贝位集装箱
堆存状态和客户提箱顺序已知为前提,据此分析模型基本假设条件,对模型变量进行描
述,构建集装箱堆场进口箱翻箱落位优选问题的 Markov 决策过程模型。
为了求解翻箱落位优选模型,设计了 ε - greedy Q 学习算法。根据堆场各因素对翻
箱率影响程度,选择关键因素对 Q 学习算法的多维状态空间进行表述,实时反映系统动
态。确定动作集合和奖惩制度,某一翻箱作业完成后,通过立即回报反馈该动作的优劣。
通过理论分析确定算法学习因子、折扣因子和探索因子随学习幕数的变化趋势,设计动
作的探索策略,平衡算法的收敛性和整体最优性。
最后设计不同规模贝位情况的算例,验证 ε - greedy Q 学习算法求解进口箱翻箱落
位优选问题的性能,实验结果表明:1)Q 学习算法求解结果相比 Kim 翻箱量估计公式,
优化率在 40%以上;2)相比参考算法 OH 算法和 IH 算法,Q 学习算法在求解大规模问
题时,二次翻箱量平均优化率分别在 50%和 10%以上;3)不同规模贝位单个算例的 Q
学习算法求解结果相对 OH、IH 算法改进稳定,100 个算例中最多有一个算例求解结果
劣于 OH;最多有四个算例劣于 IH。
关键词:集装箱运输;堆场翻箱落位优选;堆场翻箱率;Q 学习算法;Markov 决策
过程模型
万方数据
基于 Q 学习算法的集装箱堆场翻箱落位优选
- II -
Reshuffle Optimization in Container Terminal based on
Q Learning Algorithm
Abstract
Port is the most important water and land transportation hub in the integrated
transportation network, and the largest collection center in the logistics supple chain. With the
deepening of goods containerization, the container throughput has increased in the terminal, the
storage pressure of container yard is increasing, and the yard resources have become tense. In
order to improve the quality of customer service and enhance the comprehensive
competitiveness of the port, it is essential to optimize the scheduling management within the
container terminal. The reshuffle ratio of the yard is one of the most important indicators of
container terminal scheduling management. The reshuffle optimization during the picking
process can effectively reduce the relocation probability.
For the problem of reshuffle in container yard, firstly, the classification analysis is
carried out from three cases: import container reshuffle, export container reshuffle and
reshuffle when containers are moved. Then the paper analyzes the reasons for the reshuffle
from three aspects: internal factors, external factors and other irresistible factors, propose
some methods for reducing the amount of reshuffle from the aspect of yard management.
Finally, it is determined that the research object of this paper is the import containers, and it is
preferable to study the reshuffle optimization of the import containers to reduce the reshuffle
ratio of the yard.
The optimization goals of the reshuffle problem in the yard is to minimize the reshuffle
number of a bay, and the prerequisite is that the storage status of the containers in a bay and
the customers' extraction order of the containers are known. Based on this, the basic
assumptions of the model are analyzed, the model variables are described, and the Markov
decision process model for reshuffle optimization problem of the import containers in the
container yard is constructed.
In order to solve the reshuffle optimization model, a ε-greedy Q learning algorithm is
designed. According to the influence degree of each factor in the yard on the reshuffle ratio,
the key factors are selected to describe the multidimensional state space of the Q learning
algorithm to reflect the system dynamics in real time. Determine the action set and reward and
punishment system, after a certain relocation operation is completed, through the immediate
return feedback of the pros and cons of the action. Through theoretical analysis, determining
万方数据
大连理工大学专业学位硕士学位论文
- III -
the change trend of learning factor, discount factor and exploration factor with the number of
learning scenes, and designing the exploration strategy of action, in order to balance the
convergence and overall result optimality of the algorithm.
Finally, the examples of container bays with different scales is designed to verify the
performance of the ε- greedy Q learning algorithm which is to solve the problem of reshuffle
optimization of the import containers. The experimental results show that: 1) compared with
the estimation formula of the number of rehandles from Kim, the optimization rate of the Q
learning algorithm is more than 40%. 2) Compared with the reference algorithm OH
algorithm and IH algorithm, when the Q learning algorithm solves large-scale problems, the
average optimization rate of the second reshuffle is 50% and 10% respectively; 3) The
solution result of Q learning algorithm of a single case with different scales is improved
stability compared with the OH and IH algorithms. At most one of the 100 examples, the
result is inferior to OH; at most four examples are inferior to IH.
Key Words:Container Transportation; Reshuffle Optimization; Yard Reshuffle Ratio; Q
Learning Algorithm; Markov decision process model
万方数据
剩余48页未读,继续阅读
陌陌的日记
- 粉丝: 12
- 资源: 319
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功
评论0