欢迎来到冰点文库! | 帮助中心 分享价值,成长自我!
冰点文库
全部分类
  • 临时分类>
  • IT计算机>
  • 经管营销>
  • 医药卫生>
  • 自然科学>
  • 农林牧渔>
  • 人文社科>
  • 工程科技>
  • PPT模板>
  • 求职职场>
  • 解决方案>
  • 总结汇报>
  • ImageVerifierCode 换一换
    首页 冰点文库 > 资源分类 > DOCX文档下载
    分享到微信 分享到微博 分享到QQ空间

    纺织专业 人工神经网络 中英文 外文 资料 文献 原文和翻译.docx

    • 资源ID:9358029       资源大小:426.97KB        全文页数:24页
    • 资源格式: DOCX        下载积分:3金币
    快捷下载 游客一键下载
    账号登录下载
    微信登录下载
    三方登录下载: 微信开放平台登录 QQ登录
    二维码
    微信扫一扫登录
    下载资源需要3金币
    邮箱/手机:
    温馨提示:
    快捷下载时,用户名和密码都是您填写的邮箱或者手机号,方便查询和重复下载(系统自动生成)。
    如填写123,账号就是123,密码也是123。
    支付方式: 支付宝    微信支付   
    验证码:   换一换

    加入VIP,免费下载
     
    账号:
    密码:
    验证码:   换一换
      忘记密码?
        
    友情提示
    2、PDF文件下载后,可能会被浏览器默认打开,此种情况可以点击浏览器菜单,保存网页到桌面,就可以正常下载了。
    3、本站不支持迅雷下载,请使用电脑自带的IE浏览器,或者360浏览器、谷歌浏览器下载即可。
    4、本站资源下载后的文档和图纸-无水印,预览文档经过压缩,下载后原文更清晰。
    5、试题试卷类文档,如果标题没有明确说明有答案则都视为没有答案,请知晓。

    纺织专业 人工神经网络 中英文 外文 资料 文献 原文和翻译.docx

    1、纺织专业 人工神经网络 中英文 外文 资料 文献 原文和翻译Textile Research Journal Article Use of Artificial Neural Networks for Determining the LevelingAction Point at the Auto-leveling Draw FrameAssad Farooq1and Chokri CherifInstitute of Textile and Clothing Technology, Technische Universitt Dresden. Dresden, GermanyAbstract

    2、Artificial neural networks with their ability of learning from data have been successfully applied in the textile industry. The leveling action point is one of the important auto-leveling parameters of the drawing frame and strongly influences the quality of the manufactured yarn. This paper reports

    3、 a method of predicting the leveling action point using artificial neural networks. Various leveling action point affecting variables were selected as inputs for training the artificial neural networks with the aim to optimize the auto-leveling by limiting the leveling action point search range. The

    4、 Levenberg Marquardt algorithm is incorporated into the back-propagation to accelerate the training and Bayesian regularization is applied to improve the generalization of the networks. The results obtained are quite promising. Key words: artificial neural networks, auto-lev-eling, draw frame, level

    5、ing action point。 The evenness of the yarn plays an increasingly significant role in the textile industry, while the sliver evenness is one of the critical factors when producing quality yarn. The sliver evenness is also the major criteria for the assessment of the operation of the draw frame. In pr

    6、inciple, there are two approaches to reduce the sliver irregularities. One is to study the drafting mechanism and recognize the causes for irregularities, so that means may be found to reduce them. The other more valuable approach is to use auto-levelers 1, since in most cases the doubling is inadeq

    7、uate to correct the variations in sliver. The control of sliver irregularities can lower the dependence on card sliver uniformity, ambient conditions, and frame parameters. At the auto-leveler draw frame (RSB-D40) the thickness variations in the fed sliver are continually monitored by a mechanical d

    8、evice (a tongue-groove roll) and subsequently converted into electrical signals. The measured values are transmitted to an electronic memory with a variable, the time delayed response. The time delay allows the draft between the mid-roll and the delivery roll of the draw frame to adjust exactly at t

    9、hat moment when the defective sliver piece, which had been measured by a pair of scanning rollers, finds itself at a point of draft. At this point, a servo motor operates depending upon the amount of variation detected in the sliver piece. The distance that separates the scanning rollers pair and th

    10、e point of draft is called the zero point of regulation or the leveling action point (LAP) as shown in Figure 1. This leads to the calculated correction on the corresponding defective material 2,3. In auto-leveling draw frames, especially in the case of a change of fiber material, or batches the mac

    11、hine settings and process controlling parameters must be optimized. The LAP is the most important auto-leveling parameter which is influenced by various parameters such as feeding speed, material, break draft gauge, main draft gauge, feeding tension, break draft, and setting of the sliver guiding ro

    12、llers etc.Use of Artificial Neural Networks for Determining the Leveling Action Point A. Farooq and C. Cherif Figure 1 Schematic diagram of an auto-leveler drawing frame. Previously, the sliver samples had to be produced with different settings, taken to the laboratory, and examined on the evenness

    13、tester until the optimum LAP was found (manual search). Auto-leveler draw frame RSB-D40 implements an automatic search function for the optimum determination of the LAP. During this function, the sliver is automatically scanned by adjusting the different LAPs temporarily and the resulted values are

    14、recorded. During this process, the quality parameters are constantly monitored and an algorithm automatically calculates the optimum LAP by selecting the point with the minimum sliver CV%. At present a search range of 120 mm is scanned, i.e. 21 points are examined using 100 m of sliver in each case;

    15、 therefore 2100 m of sliver is necessary to carry out the search function. This is a very time-consuming method accompanied by the material and production losses, and hence directly affecting the cost parameters. In this work, we have tried to find out the possibility of predicting the LAP, using ar

    16、tificial neural net-works, to limit the automatic search span and to reduce the above-mentioned disadvantages.Artificial Neural NetworksThe motivation of using artificial neural networks lies in their flexibility and power of information processing that conventional computing methods do not have. Th

    17、e neural network system can solve a problem “by experience and learning” the inputoutput patterns provided by the user. In the field of textiles, artificial neural networks (mostly using back-propagation) have been extensively studied during the last two decades 46. In the field of spinning previous

    18、 research has concentrated on predicting the yarn properties and the spinning process performance using the fiber properties or a combination of fiber properties and machine settings as the input of neural networks 712.Back-propagation is a supervised learning technique most frequently used for arti

    19、ficial neural network training. The back-propagation algorithm is based on the Widrow-Hoff delta learning rule in which the weight adjustment is carried out through the mean square error of the output response to the sample input 13. The set of these sample patterns is repeatedly presented to the ne

    20、twork until the error value is minimized. The back-propagation algorithm uses the steepest descent method, which is essentially a first-order method to determine a suitable direction of gradient movement.OverfittingThe goal of neural network training is to produce a network which produces small erro

    21、rs on the training set, and which also responds properly to novel inputs. When a network performs as well on novel inputs as on training set inputs, the network is said to be well generalized. The generalization capacity of the network is largely governed by the network architecture (number of hidde

    22、n neurons) and this plays a vital role during the training. A network which is not complex enough to learn all the information in the data is said to be underfitted, while a network that is too complex to fit the “noise” in the data leads to overfitting. “Noise” means variation in the target values

    23、that are unpredictable from the inputs of a specific network. All standard neural network architectures such as the fully connected multi-layer perceptron are prone to overfitting. Moreover, it is very difficult to acquire the noise free data from the spinning industry due to the dependence of end p

    24、roducts on the inherent material variations and environmental conditions, etc. Early stopping is the most commonly used technique to tackle this problem. This involves the division of training data into three sets, i.e. a training set, a validation set and a test set, with the drawback that a large

    25、part of the data (validation set) can never be the part of the training.Regularization The other solution of overfitting is regularization, which is the method of improving the generalization by constraining the size of the network weights. Mackay 14 discussed a practical Bayesian framework for back

    26、-propagation networks, which consistently produced networks with good generalization. The initial objective of the training process is to mini-mize the sum of square errors: (1)Where are the targets and are the neural network responses to the respective targets. Typically, training aims to reduce th

    27、e sum of squared errors F=Ed. However, regularization adds an additional term, the objective function, (2)In equation (2), is the sum of squares of the network weights, and and are objective function parameters. The relative size of the objective function parameters dictates the emphasis for trainin

    28、g. If , training will emphasize weight size reduction at the expense of network errors, thus producing a smoother network response 15.The Bayesian School of statistics is based on a different view of what it means to learn from data, in which probability is used to represent the uncertainty about th

    29、e relationship being learned. Before seeing any data, the prior opinions about what the true relationship might be can be expressed in a probability distribution over the network weights that define this relationship. After the program conceives the data, the revised opinions are captured by a poste

    30、rior distribution over network weights. Network weights that seemed plausible before, but which do not match the data very well, will now be seen as being much less likely, whilethe probability for values of the weights that do fit the data well will have increased 16. In the Bayesian framework the

    31、weights of the network are considered random variables. After the data is taken, the posterior probability function for the weights can be updated according to Bayes rule: (3)In equation (3), D represents the data set, M is the particular neural network model used, and w is the vector of network wei

    32、ghts. is the prior probability, which represents our knowledge of the weights before any data is collected. is the likelihood function, which is the probability of data occurring, given the weights w.is a normalization factor, which guarantees that the total probability is 1 15. In this study, we employed the MATLAB Neural Net-works Toolbox function “trainbr” which is an incorporation of the LevenbergMarqaurdt algorithm and the Bayesian regularization theorem (or Bayesian learning) into back-propagation t


    注意事项

    本文(纺织专业 人工神经网络 中英文 外文 资料 文献 原文和翻译.docx)为本站会员主动上传,冰点文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知冰点文库(点击联系客服),我们立即给予删除!

    温馨提示:如果因为网速或其他原因下载失败请重新下载,重复下载不扣分。




    关于我们 - 网站声明 - 网站地图 - 资源地图 - 友情链接 - 网站客服 - 联系我们

    copyright@ 2008-2023 冰点文库 网站版权所有

    经营许可证编号:鄂ICP备19020893号-2


    收起
    展开