日本黄色一级经典视频|伊人久久精品视频|亚洲黄色色周成人视频九九九|av免费网址黄色小短片|黄色Av无码亚洲成年人|亚洲1区2区3区无码|真人黄片免费观看|无码一级小说欧美日免费三级|日韩中文字幕91在线看|精品久久久无码中文字幕边打电话

當前位置:首頁 > 醫(yī)療電子 > 醫(yī)療電子
[導讀] 很少有人知道當手臂拿起一個球時神經(jīng)、臂膀和傳感系統(tǒng)之間的交互。為了模擬這一自然反應過程,可以通過微處理器、嵌入式控制軟件、執(zhí)行機構和傳感器來構造這一系統(tǒng)從而來研究它們之間的復雜關系。這也是美國國

      很少有人知道當手臂拿起一個球時神經(jīng)、臂膀和傳感系統(tǒng)之間的交互。為了模擬這一自然反應過程,可以通過微處理器、嵌入式控制軟件、執(zhí)行機構和傳感器來構造這一系統(tǒng)從而來研究它們之間的復雜關系。這也是美國國防高級研究計劃署(DARPA)革命性假肢計劃所面臨的挑戰(zhàn)。

      美國約翰霍普金斯大學應用物理實驗室是領導性的全球團隊,包括政府機構、大學、私有企業(yè),他們的任務是開發(fā)世界上最先進的假肢,此假肢由神經(jīng)輸入控制,使佩戴者感覺是一個真的手臂一樣能夠以一定的速度、靈敏度和力去運動。先進的傳感反饋技術能夠感知物理輸入,如壓力、力和溫度。

      這個項目中具有里程碑意義的關鍵部分是虛擬綜合環(huán)境的開發(fā),一個完整的手臂系統(tǒng)的仿真環(huán)境使用The Mathworks工具和基于模型設計。虛擬綜合環(huán)境具有標準化的架構和定義完善的界面,能夠使二十多不同領域?qū)<液芎玫睾献鳌?/p>

      The Mathworks工具基于模型設計也被用在其他開發(fā)階段,包括對臂的機械系統(tǒng)進行建模、測試新的神經(jīng)解碼算法和開發(fā)與驗證控制算法。

 

      為 DARPA計劃開發(fā)的兩個原型手臂使用了目標肌肉神經(jīng)系統(tǒng),這項技術是由芝加哥康復研究院Todd Kuiken博士研發(fā)的,內(nèi)容包括從被切除手臂到未使用的傷害處的肌肉區(qū)域的殘留神經(jīng)的傳輸。在臨床評估中,第一個原型能夠使患者完成各種功能任務,包括從口袋里拿一個信用卡。 

Virtual Integration Environment Architecture

The VIE architecture consists of five main modules: Input, Signal Analysis, Controls, Plant, and Presentation.

The Input module comprises all the input devices that patients can use to signal their intent, including surface electromyograms (EMGs), cortical and peripheral nerve implants, implantable myoelectric sensors (IMESs) and more conventional digital and analog inputs for switches, joysticks, and other control sources used by clinicians. The Signal Analysis module performs signal processing and filtering. More important, this module applies pattern recognition algorithms that interpret raw input signals to extract the user’s intent and communicate that intent to the Controls module. In the Controls module, those commands are mapped to motor signals that control the individual motors that actuate the limb, hand, and fingers.

The Plant module consists of a physical model of the limb’s mechanics. The Presentation module produces a three-dimensional (3D) rendering of the arm’s movement (Figure 1).


圖1 假肢三維視圖

Interfacing with the Nervous System

Simulink® and the VIE were essential to developing an interface to the nervous system that allows natural and intuitive control of the prosthetic limb system. Researchers record data from neural device implants while the subjects perform tasks such as reaching for a ball in the virtual environment. The VIE modular input systems receive this data, and MATLAB® algorithms decode the subject’s intent by using pattern recognition to correlate neural activity with the subject’s movement (Figure 2). The results are integrated back into the VIE, where experiments can be run in real time.


 
圖2 紐布朗斯威克大學開發(fā)了MATLAB應用程序,記錄用于模式識別的運動數(shù)據(jù)。

The same workflow has been used to develop input devices of all kinds, some of which are already being tested by prosthetic limb users at the Rehabilitation Institute of Chicago.

Building Real-Time Prototype Controllers

The Signal Analysis and Controls modules of the VIE form the heart of the control system that will ultimately be deployed in the prosthetic arm. At APL, we developed the software for these modules. Individual algorithms were developed in MATLAB using the Embedded MATLAB™ subset and then integrated into a Simulink model of the system as function blocks. To create a real-time prototype of the control system, we generated code for the complete system, including the Simulink and Embedded MATLAB components, with Real-Time Workshop®, and deployed this code to xPC Target™.

This approach brought many advantages. Using Model-Based Design and Simulink, we modeled the complete system and simulated it to optimize and verify the design. We were able to rapidly build and test a virtual prototype system before committing to a specific hardware platform. With Real-Time Workshop Embedded Coder™ we generated target-specific code for our processor. Because the code is generated from a Simulink system model that has been safety-tested and verified through simulation, there is no hand-coding step that could introduce errors or unplanned behaviors. As a result, we have a high degree of confidence that the Modular Prosthetic Limb will perform as intended and designed.

Physical Modeling and Visualization

To perform closed-loop simulations of our control system, we developed a plant model representing the inertial properties of the limb system. We began with CAD assemblies of limb components designed in SolidWorks® by our partners. We used the CAD assemblies to automatically generate a SimMechanics™ model of the limb linked to our control system in Simulink.

Finally, we linked the plant model to a Java™ 3D rendering engine developed at the University of Southern California to show a virtual limb moving in a simulated environment.

Clinical Application

Given the powerful virtual system framework, we were also able to create a useful and intuitive clinical environment for system configuration and training. Clinicians can configure parameters in the VIE and manage test sessions with volunteer subjects using a GUI that we created in MATLAB (Figure 3).

Clinicians interact with this application on a host PC that communicates with the xPC Target system running the control software in real time. A third PC is used for 3D rendering and display of the virtual limb. During tests of actual limbs, we can correlate and visualize control signals while the subject is moving.

Looking Ahead

Using Model-Based Design, the Revolutionizing Prosthetics team has delivered Proto 1, Proto 2, and the first version of the VIE ahead of schedule. Currently we are in the process of developing a detailed design of the Modular Prosthetic Limb, the version that we will deliver to DARPA.

Many of our partner institutions use the VIE as a test bed as they continue to improve their systems, and we envision the VIE continuing as a platform for further development in prosthetics and neuroscience for years to come. Our team has established a development process that we can use to rapidly assemble systems from reusable models and implement on prototype hardware, not only for the Revolutionizing Prosthetics project but for related programs as well.

As we meet the challenge of building a mechatronic system that mimics natural motion, we strive to match the perseverance and commitment that our volunteer subjects and the amputee population at large demonstrate every day.

Approved for Public Release, Distribution Unlimited.

Mimicking Nature on a Deadline

Developing a mechatronic system that replicates natural motion and preparing it for clinical trials in just four years, as mandated by DARPA, requires breakthroughs in neural control, sensory input, advanced mechanics and actuators, and prosthesis design.

State-of-the-art prosthetic arms today typically have just three active degrees of freedom: elbow flex/extend, wrist rotate, and grip open/close. Proto 1, our first prototype, added five more degrees of freedom, including two active degrees of freedom at the shoulder (flexion/extension and internal/external rotation), wrist flexion/extention, and additional hand grips. To emulate natural movement, we needed to go far beyond the advances in Proto 1.

Proto 2, which was developed as an electromechanical proof of concept, had more than 22 degrees of freedom, including additional side-to-side movements at the shoulder (abduction/adduction), wrist (radial/unlar deviation), and independent articulation of the fingers. The hand can also be commanded into multiple highly functional coordinated “grasps.”

The Modular Prosthetic Limb—the version that we will deliver to DARPA—will have 27 degrees of freedom, as well as the ability to sense temperature, contact, pressure, and vibration.

 

Proto 2 hand grasps. Click on image to see enlarged view.

Products Used

MATLAB®
Real-Time Workshop® 
Real-Time Workshop® Embedded Coder™ 
SimMechanics™ 
Simulink® 
xPC Target
Resources

Johns Hopkins University Applied Physics Laboratory
Model-Based Design

本站聲明: 本文章由作者或相關機構授權發(fā)布,目的在于傳遞更多信息,并不代表本站贊同其觀點,本站亦不保證或承諾內(nèi)容真實性等。需要轉(zhuǎn)載請聯(lián)系該專欄作者,如若文章內(nèi)容侵犯您的權益,請及時聯(lián)系本站刪除。
換一批
延伸閱讀

舍弗勒以"專注驅(qū)動技術的科技公司"為主題亮相IAA MOBILITY 2025(B3館B40展臺) 合并緯湃科技后首次亮相IAA MOBILITY,展示拓展后的汽車產(chǎn)品組合 憑借在軟件、...

關鍵字: 電氣 軟件 驅(qū)動技術 BSP

香港2025年 9月12日 /美通社/ -- 全球領先的互聯(lián)網(wǎng)社區(qū)創(chuàng)建者 - 網(wǎng)龍網(wǎng)絡控股有限公司 ("網(wǎng)龍"或"本公司",香港交易所股票代碼:777)欣然宣布,其子公司My...

關鍵字: AI 遠程控制 控制技術 BSP

慕尼黑2025年9月12日 /美通社/ -- 慕尼黑當?shù)貢r間9月10日,在2025德國國際汽車及智慧出行博覽會(IAA MOBILITY)上,國際獨立第三方檢測、檢驗和認證機...

關鍵字: 測試 慕尼黑 模型 HUBER

深圳2025年9月11日 /美通社/ -- 2025 年 9 月 10 日,第 26 屆中國國際光電博覽會(簡稱 "CIOE 中國光博會")在深圳盛大開幕。本屆展會吸引力再創(chuàng)新高,全球超3800家優(yōu)質(zhì)...

關鍵字: 自動化 光電 CIO BSP

天津2025年9月11日 /美通社/ -- 國際能源署(IEA)數(shù)據(jù)顯示,2024 年全球數(shù)據(jù)中心電力消耗達 415 太瓦時,占全球總用電量的 1.5%,預計到 2030 年,這一數(shù)字將飆升至 945 太瓦時,近乎翻番,...

關鍵字: 模型 AI 數(shù)據(jù)中心 BSP

北京2025年9月11日 /美通社/ -- 國際9月11日上午,2025年中國國際服務貿(mào)易交易會(以下簡稱"服貿(mào)會")—體育賽事經(jīng)濟高質(zhì)量發(fā)展大會現(xiàn)場,北京經(jīng)濟技術開發(fā)區(qū)工委委員、管委會副主...

關鍵字: 5G BSP GROUP MOTOR

柏林2025年9月9日 /美通社/ -- 2025年9月5日,納斯達克上市公司優(yōu)克聯(lián)集團(NASDAQ: UCL)旗下全球互聯(lián)品牌GlocalMe,正式亮相柏林國際消費電子展(IFA 2025),重磅推出融合企...

關鍵字: LOCAL LM BSP 移動網(wǎng)絡

深圳2025年9月9日 /美通社/ -- PART 01活動背景 當技術的鋒芒刺穿行業(yè)壁壘,萬物互聯(lián)的生態(tài)正重塑產(chǎn)業(yè)疆域。2025年,物聯(lián)網(wǎng)產(chǎn)業(yè)邁入?"破界創(chuàng)造"與"共生進化"?的裂變時代——AI大模型消融感知邊界,...

關鍵字: BSP 模型 微信 AIOT

"出海無界 商機無限"助力企業(yè)構建全球競爭力 深圳2025年9月9日 /美通社/ -- 2025年8月28日, 由領先商業(yè)管理媒體世界經(jīng)理人攜手環(huán)球資源聯(lián)合主辦、深圳?前海出海e站通協(xié)辦的...

關鍵字: 解碼 供應鏈 AI BSP

柏林2025年9月9日 /美通社/ -- 柏林當?shù)貢r間9月6日,在2025德國柏林國際電子消費品展覽會(International Funkausstellung...

關鍵字: 掃地機器人 耳機 PEN BSP
關閉