mirror of
https://github.com/dptech-corp/Uni-Lab-OS.git
synced 2026-02-05 22:15:04 +00:00
Compare commits
417 Commits
main
...
fd73bb7dcb
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
fd73bb7dcb | ||
|
|
a02cecfd18 | ||
|
|
d6accc3f1c | ||
|
|
39dc443399 | ||
|
|
37b1fca962 | ||
|
|
216f19fb62 | ||
|
|
ec7ca6a1fe | ||
|
|
4c8022ee95 | ||
|
|
ad21644db0 | ||
|
|
9dfd58e9af | ||
|
|
31c9f9a172 | ||
|
|
02cd8de4c5 | ||
|
|
a66603ec1c | ||
|
|
ec015e16cd | ||
|
|
965bf36e8d | ||
|
|
aacf3497e0 | ||
|
|
657f952e7a | ||
|
|
0165590290 | ||
|
|
daea1ab54d | ||
|
|
93cb307396 | ||
|
|
1c312772ae | ||
|
|
bad1db5094 | ||
|
|
f26eb69eca | ||
|
|
12c0770c92 | ||
|
|
3d2d428a96 | ||
|
|
78bf57f590 | ||
|
|
e227cddab3 | ||
|
|
f2b993643f | ||
|
|
2e14bf197c | ||
|
|
66c18c080a | ||
|
|
a1c34f138e | ||
|
|
75bb5ec553 | ||
|
|
bb95c89829 | ||
|
|
394c140830 | ||
|
|
e6d8d41183 | ||
|
|
847a300af3 | ||
|
|
a201d7c307 | ||
|
|
3433766bc5 | ||
|
|
7e9e93b29c | ||
|
|
9e1e6da505 | ||
|
|
8a0f000bab | ||
|
|
2ffeb49acb | ||
|
|
5fec753fb9 | ||
|
|
acbaff7bb7 | ||
|
|
706323dc3e | ||
|
|
b0804d939c | ||
|
|
97788b4e07 | ||
|
|
39cc280c91 | ||
|
|
d0ac452405 | ||
|
|
152d3a7563 | ||
|
|
ef14737839 | ||
|
|
5d5569121c | ||
|
|
d23e85ade4 | ||
|
|
02afafd423 | ||
|
|
6ac510dcd2 | ||
|
|
ed56c1eba2 | ||
|
|
16ee3de086 | ||
|
|
ced961050d | ||
|
|
11b2c99836 | ||
|
|
04024bc8a3 | ||
|
|
154048107d | ||
|
|
0b896870ba | ||
|
|
ee609e4aa2 | ||
|
|
5551fbf360 | ||
|
|
e13b250632 | ||
|
|
b8278c5026 | ||
|
|
53e767a054 | ||
|
|
cf7032fa81 | ||
|
|
97681ba433 | ||
|
|
3fa81ab4f6 | ||
|
|
9f4a69ddf5 | ||
|
|
05ae4e72df | ||
|
|
2870c04086 | ||
|
|
343e87df0d | ||
|
|
5d0807cba6 | ||
|
|
4875977d5f | ||
|
|
956b1c905b | ||
|
|
944911c52a | ||
|
|
a13b790926 | ||
|
|
9feadd68c6 | ||
|
|
c68d5246d0 | ||
|
|
49073f2c77 | ||
|
|
b2afc29f15 | ||
|
|
4061280f6b | ||
|
|
6a681e1d73 | ||
|
|
653e6e1ac3 | ||
|
|
2c774bcd1d | ||
|
|
2ba395b681 | ||
|
|
b6b3d59083 | ||
|
|
f40e3f521c | ||
|
|
7cc2fe036f | ||
|
|
f81d20bb1d | ||
|
|
db1b5a869f | ||
|
|
0136630700 | ||
|
|
3c31811f9e | ||
|
|
64f02ff129 | ||
|
|
7d097b8222 | ||
|
|
d266d21104 | ||
|
|
b6d0bbcb17 | ||
|
|
31ebff8e37 | ||
|
|
2132895ba2 | ||
|
|
850eeae55a | ||
|
|
d869c14233 | ||
|
|
24101b3cec | ||
|
|
3bf8aad4d5 | ||
|
|
a599eb70e5 | ||
|
|
0bf6994f95 | ||
|
|
c36f53791c | ||
|
|
eb4d2d96c5 | ||
|
|
8233c41b1d | ||
|
|
0dfd4ce8a8 | ||
|
|
7953b3820e | ||
|
|
eed233fa76 | ||
|
|
0c55147ee4 | ||
|
|
ce6267b8e0 | ||
|
|
975e51cd96 | ||
|
|
c5056b381c | ||
|
|
c35da65b15 | ||
|
|
659cf05be6 | ||
|
|
3b8deb4d1d | ||
|
|
c796615f9f | ||
|
|
a5bad6074f | ||
|
|
1d3a07a736 | ||
|
|
cc2cd57cdf | ||
|
|
39bb7dc627 | ||
|
|
0fda155f55 | ||
|
|
6e3eacd2f0 | ||
|
|
062f1a2153 | ||
|
|
61e8d67800 | ||
|
|
d0884cdbd8 | ||
|
|
545ea45024 | ||
|
|
b9ddee8f2c | ||
|
|
a0c5095304 | ||
|
|
e504505137 | ||
|
|
4d9d5701e9 | ||
|
|
6016c4b588 | ||
|
|
be02bef9c4 | ||
|
|
e62f0c2585 | ||
|
|
b6de0623e2 | ||
|
|
9d081e9fcd | ||
|
|
85a58e3464 | ||
|
|
85590672d8 | ||
|
|
1d4018196d | ||
|
|
5d34f742af | ||
|
|
5bef19e6d6 | ||
|
|
f816799753 | ||
|
|
a45d841769 | ||
|
|
7f0b33b3e3 | ||
|
|
2006406a24 | ||
|
|
f94985632b | ||
|
|
12ba110569 | ||
|
|
97212be8b7 | ||
|
|
9bdd42f12f | ||
|
|
627140da03 | ||
|
|
5ceedb0565 | ||
|
|
8c77a20c43 | ||
|
|
3ff894feee | ||
|
|
fa5896ffdb | ||
|
|
eb504803ac | ||
|
|
8b0c845661 | ||
|
|
693873bfa9 | ||
|
|
57da2d8da2 | ||
|
|
8d1fd01259 | ||
|
|
388259e64b | ||
|
|
2c130e7f37 | ||
|
|
9f7c3f02f9 | ||
|
|
19dd80dcdb | ||
|
|
9d5ed627a2 | ||
|
|
2d0ff87bc8 | ||
|
|
d78475de9a | ||
|
|
88ae56806c | ||
|
|
95dd8beb81 | ||
|
|
4ab3fadbec | ||
|
|
229888f834 | ||
|
|
b443b39ebf | ||
|
|
0434bbc15b | ||
|
|
5791b81954 | ||
|
|
bd51c74fab | ||
|
|
ba81cbddf8 | ||
|
|
4e92a26057 | ||
|
|
c2895bb197 | ||
|
|
0423f4f452 | ||
|
|
41390fbef9 | ||
|
|
98bdb4e7e4 | ||
|
|
30037a077a | ||
|
|
6972680099 | ||
|
|
9d2c93807d | ||
|
|
e728007bc5 | ||
|
|
9c5ecda7cc | ||
|
|
2d26c3fac6 | ||
|
|
f5753afb7c | ||
|
|
398b2dde3f | ||
|
|
62c4135938 | ||
|
|
027b4269c4 | ||
|
|
3757bd9c58 | ||
|
|
c75b7d5aae | ||
|
|
dfc635189c | ||
|
|
d8f3ebac15 | ||
|
|
4a1e703a3a | ||
|
|
55d22a7c29 | ||
|
|
03a4e4ecba | ||
|
|
2316c34cb5 | ||
|
|
a8887161d3 | ||
|
|
25834f5ba0 | ||
|
|
a1e9332b51 | ||
|
|
357fc038ef | ||
|
|
fd58ef07f3 | ||
|
|
93dee2c1dc | ||
|
|
70fbf19009 | ||
|
|
9149155232 | ||
|
|
1ca1792e3c | ||
|
|
485e7e8dd2 | ||
|
|
4ddabdcb65 | ||
|
|
a5b0325301 | ||
|
|
50b44938c7 | ||
|
|
df0d2235b0 | ||
|
|
4e434eeb97 | ||
|
|
ca027bf0eb | ||
|
|
635a332b4e | ||
|
|
edf7a117ca | ||
|
|
70b2715996 | ||
|
|
7e8dfc2dc5 | ||
|
|
9b626489a8 | ||
|
|
03fe208743 | ||
|
|
e913e540a3 | ||
|
|
aed39b648d | ||
|
|
8c8359fab3 | ||
|
|
5d20be0762 | ||
|
|
09f745d300 | ||
|
|
bbcbcde9a4 | ||
|
|
42b437cdea | ||
|
|
ffd0f2d26a | ||
|
|
32422c0b3d | ||
|
|
c44e597dc0 | ||
|
|
4eef012a8e | ||
|
|
ac69452f3c | ||
|
|
57b30f627b | ||
|
|
2d2a4ca067 | ||
|
|
a2613aad4c | ||
|
|
54f75183ff | ||
|
|
735be067dc | ||
|
|
0fe62d64f0 | ||
|
|
2d4ecec1e1 | ||
|
|
0f976a1874 | ||
|
|
b263a7e679 | ||
|
|
7c7f1b31c5 | ||
|
|
00e668e140 | ||
|
|
4989f65a0b | ||
|
|
9fa3688196 | ||
|
|
40fb1ea49c | ||
|
|
18b0bb397e | ||
|
|
65abc5dbf7 | ||
|
|
2455ca15ba | ||
|
|
05a3ff607a | ||
|
|
ec882df36d | ||
|
|
43b992e3eb | ||
|
|
6422fa5a9a | ||
|
|
434b9e98e0 | ||
|
|
040073f430 | ||
|
|
3d95c9896a | ||
|
|
9aa97ed01e | ||
|
|
0b8bdf5e0a | ||
|
|
299f010754 | ||
|
|
15ce0d6883 | ||
|
|
dec474e1a7 | ||
|
|
5f187899fc | ||
|
|
c8d16c7024 | ||
|
|
25d46dc9d5 | ||
|
|
88c4d1a9d1 | ||
|
|
81fd8291c5 | ||
|
|
3a11eb90d4 | ||
|
|
387866b9c9 | ||
|
|
7f40f141f6 | ||
|
|
6fc7ed1b88 | ||
|
|
93f0e08d75 | ||
|
|
4b43734b55 | ||
|
|
174b1914d4 | ||
|
|
704e13f030 | ||
|
|
0c42d60cf2 | ||
|
|
df33e1a214 | ||
|
|
1f49924966 | ||
|
|
609b6006e8 | ||
|
|
67c01271b7 | ||
|
|
a1783f489e | ||
|
|
a8f6527de9 | ||
|
|
54cfaf15f3 | ||
|
|
5610c28b67 | ||
|
|
cfc1ee6e79 | ||
|
|
1c9d2ee98a | ||
|
|
3fe8f4ca44 | ||
|
|
2476821dcc | ||
|
|
7b426ed5ae | ||
|
|
9bbae96447 | ||
|
|
10aabb7592 | ||
|
|
709eb0d91c | ||
|
|
14b7d52825 | ||
|
|
a5397ffe12 | ||
|
|
c6c2da69ba | ||
|
|
622e579063 | ||
|
|
196e0f7e2b | ||
|
|
a632fd495e | ||
|
|
a8cc02a126 | ||
|
|
ad2e1432c6 | ||
|
|
c3b9583eac | ||
|
|
5c47cd0c8a | ||
|
|
63ab1af45d | ||
|
|
a8419dc0c3 | ||
|
|
34f05f2e25 | ||
|
|
0dc2488f02 | ||
|
|
f13156e792 | ||
|
|
13fd1ac572 | ||
|
|
f8ef6e0686 | ||
|
|
94a7b8aaca | ||
|
|
301bea639e | ||
|
|
4b5a83efa4 | ||
|
|
2889e9be2c | ||
|
|
304aebbba7 | ||
|
|
091c9fa247 | ||
|
|
67ca45a240 | ||
|
|
7aab2ea493 | ||
|
|
62f3a6d696 | ||
|
|
eb70ad0e18 | ||
|
|
768f43880e | ||
|
|
762c3c737c | ||
|
|
ace98a4472 | ||
|
|
41eaa88c6f | ||
|
|
a1a55a2c0a | ||
|
|
2eaa0ca729 | ||
|
|
6f8f070f40 | ||
|
|
da4bd927e0 | ||
|
|
01f8816597 | ||
|
|
e5006285df | ||
|
|
573c724a5c | ||
|
|
09549d2839 | ||
|
|
50c7777cea | ||
|
|
4888f02c09 | ||
|
|
779c9693d9 | ||
|
|
ffa841a41a | ||
|
|
fc669f09f8 | ||
|
|
2ca0311de6 | ||
|
|
94cdcbf24e | ||
|
|
1cd07915e7 | ||
|
|
b600fc666d | ||
|
|
9e214c56c1 | ||
|
|
bdf27a7e82 | ||
|
|
2493fb9f94 | ||
|
|
c7a0ff67a9 | ||
|
|
711a7c65fa | ||
|
|
cde7956896 | ||
|
|
95b6fd0451 | ||
|
|
513e848d89 | ||
|
|
58d1cc4720 | ||
|
|
5676dd6589 | ||
|
|
1ae274a833 | ||
|
|
22b88c8441 | ||
|
|
81bcc1907d | ||
|
|
8cffd3dc21 | ||
|
|
a722636938 | ||
|
|
f68340d932 | ||
|
|
361eae2f6d | ||
|
|
c25283ae04 | ||
|
|
961752fb0d | ||
|
|
55165024dd | ||
|
|
6ddceb8393 | ||
|
|
4e52c7d2f4 | ||
|
|
0b56efc89d | ||
|
|
a27b93396a | ||
|
|
2a60a6c27e | ||
|
|
5dda94044d | ||
|
|
0cfc6f45e3 | ||
|
|
831f4549f9 | ||
|
|
f4d4eb06d3 | ||
|
|
e3b8164f6b | ||
|
|
78c04acc2e | ||
|
|
cd0428ea78 | ||
|
|
bdddbd57ba | ||
|
|
a312de08a5 | ||
|
|
68513b5745 | ||
|
|
19027350fb | ||
|
|
bbbdb06bbc | ||
|
|
cd84e26126 | ||
|
|
ce5bab3af1 | ||
|
|
82d9ef6bf7 | ||
|
|
332b33c6f4 | ||
|
|
1ec642ee3a | ||
|
|
7d8e6d029b | ||
|
|
5ec8a57a1f | ||
|
|
ae3c1100ae | ||
|
|
14bc2e6cda | ||
|
|
9f823a4198 | ||
|
|
02c79363c1 | ||
|
|
227ff1284a | ||
|
|
4b7bde6be5 | ||
|
|
8a669ac35a | ||
|
|
a1538da39e | ||
|
|
0063df4cf3 | ||
|
|
e570ba4976 | ||
|
|
e8c1f76dbb | ||
|
|
f791c1a342 | ||
|
|
ea60cbe891 | ||
|
|
eac9b8ab3d | ||
|
|
573bcf1a6c | ||
|
|
50e93cb1af | ||
|
|
fe1a029a9b | ||
|
|
662c063f50 | ||
|
|
01cbbba0b3 | ||
|
|
e6c556cf19 | ||
|
|
0605f305ed | ||
|
|
37d8108ec4 | ||
|
|
6081dac561 | ||
|
|
5b2d066127 | ||
|
|
06e66765e7 | ||
|
|
98ce360088 | ||
|
|
5cd0f72fbd | ||
|
|
343f394203 | ||
|
|
46aa7a7bd2 | ||
|
|
a66369e2c3 |
@@ -1,60 +0,0 @@
|
|||||||
# unilabos: Production package (depends on unilabos-env + pip unilabos)
|
|
||||||
# For production deployment
|
|
||||||
|
|
||||||
package:
|
|
||||||
name: unilabos
|
|
||||||
version: 0.10.17
|
|
||||||
|
|
||||||
source:
|
|
||||||
path: ../../unilabos
|
|
||||||
target_directory: unilabos
|
|
||||||
|
|
||||||
build:
|
|
||||||
python:
|
|
||||||
entry_points:
|
|
||||||
- unilab = unilabos.app.main:main
|
|
||||||
script:
|
|
||||||
- set PIP_NO_INDEX=
|
|
||||||
- if: win
|
|
||||||
then:
|
|
||||||
- copy %RECIPE_DIR%\..\..\MANIFEST.in %SRC_DIR%
|
|
||||||
- copy %RECIPE_DIR%\..\..\setup.cfg %SRC_DIR%
|
|
||||||
- copy %RECIPE_DIR%\..\..\setup.py %SRC_DIR%
|
|
||||||
- pip install %SRC_DIR%
|
|
||||||
- if: unix
|
|
||||||
then:
|
|
||||||
- cp $RECIPE_DIR/../../MANIFEST.in $SRC_DIR
|
|
||||||
- cp $RECIPE_DIR/../../setup.cfg $SRC_DIR
|
|
||||||
- cp $RECIPE_DIR/../../setup.py $SRC_DIR
|
|
||||||
- pip install $SRC_DIR
|
|
||||||
|
|
||||||
requirements:
|
|
||||||
host:
|
|
||||||
- python ==3.11.14
|
|
||||||
- pip
|
|
||||||
- setuptools
|
|
||||||
- zstd
|
|
||||||
- zstandard
|
|
||||||
run:
|
|
||||||
- zstd
|
|
||||||
- zstandard
|
|
||||||
- networkx
|
|
||||||
- typing_extensions
|
|
||||||
- websockets
|
|
||||||
- pint
|
|
||||||
- fastapi
|
|
||||||
- jinja2
|
|
||||||
- requests
|
|
||||||
- uvicorn
|
|
||||||
- opcua
|
|
||||||
- pyserial
|
|
||||||
- pandas
|
|
||||||
- pymodbus
|
|
||||||
- matplotlib
|
|
||||||
- pylibftdi
|
|
||||||
- uni-lab::unilabos-env ==0.10.17
|
|
||||||
|
|
||||||
about:
|
|
||||||
repository: https://github.com/deepmodeling/Uni-Lab-OS
|
|
||||||
license: GPL-3.0-only
|
|
||||||
description: "UniLabOS - Production package with minimal ROS2 dependencies"
|
|
||||||
@@ -1,39 +0,0 @@
|
|||||||
# unilabos-env: conda environment dependencies (ROS2 + conda packages)
|
|
||||||
|
|
||||||
package:
|
|
||||||
name: unilabos-env
|
|
||||||
version: 0.10.17
|
|
||||||
|
|
||||||
build:
|
|
||||||
noarch: generic
|
|
||||||
|
|
||||||
requirements:
|
|
||||||
run:
|
|
||||||
# Python
|
|
||||||
- zstd
|
|
||||||
- zstandard
|
|
||||||
- conda-forge::python ==3.11.14
|
|
||||||
- conda-forge::opencv
|
|
||||||
# ROS2 dependencies (from ci-check.yml)
|
|
||||||
- robostack-staging::ros-humble-ros-core
|
|
||||||
- robostack-staging::ros-humble-action-msgs
|
|
||||||
- robostack-staging::ros-humble-std-msgs
|
|
||||||
- robostack-staging::ros-humble-geometry-msgs
|
|
||||||
- robostack-staging::ros-humble-control-msgs
|
|
||||||
- robostack-staging::ros-humble-nav2-msgs
|
|
||||||
- robostack-staging::ros-humble-cv-bridge
|
|
||||||
- robostack-staging::ros-humble-vision-opencv
|
|
||||||
- robostack-staging::ros-humble-tf-transformations
|
|
||||||
- robostack-staging::ros-humble-moveit-msgs
|
|
||||||
- robostack-staging::ros-humble-tf2-ros
|
|
||||||
- robostack-staging::ros-humble-tf2-ros-py
|
|
||||||
- conda-forge::transforms3d
|
|
||||||
- conda-forge::uv
|
|
||||||
|
|
||||||
# UniLabOS custom messages
|
|
||||||
- uni-lab::ros-humble-unilabos-msgs
|
|
||||||
|
|
||||||
about:
|
|
||||||
repository: https://github.com/deepmodeling/Uni-Lab-OS
|
|
||||||
license: GPL-3.0-only
|
|
||||||
description: "UniLabOS Environment - ROS2 and conda dependencies"
|
|
||||||
@@ -1,42 +0,0 @@
|
|||||||
# unilabos-full: Full package with all features
|
|
||||||
# Depends on unilabos + complete ROS2 desktop + dev tools
|
|
||||||
|
|
||||||
package:
|
|
||||||
name: unilabos-full
|
|
||||||
version: 0.10.17
|
|
||||||
|
|
||||||
build:
|
|
||||||
noarch: generic
|
|
||||||
|
|
||||||
requirements:
|
|
||||||
run:
|
|
||||||
# Base unilabos package (includes unilabos-env)
|
|
||||||
- uni-lab::unilabos ==0.10.17
|
|
||||||
# Documentation tools
|
|
||||||
- sphinx
|
|
||||||
- sphinx_rtd_theme
|
|
||||||
# Web UI
|
|
||||||
- gradio
|
|
||||||
- flask
|
|
||||||
# Interactive development
|
|
||||||
- ipython
|
|
||||||
- jupyter
|
|
||||||
- jupyros
|
|
||||||
- colcon-common-extensions
|
|
||||||
# ROS2 full desktop (includes rviz2, gazebo, etc.)
|
|
||||||
- robostack-staging::ros-humble-desktop-full
|
|
||||||
# Navigation and motion control
|
|
||||||
- ros-humble-navigation2
|
|
||||||
- ros-humble-ros2-control
|
|
||||||
- ros-humble-robot-state-publisher
|
|
||||||
- ros-humble-joint-state-publisher
|
|
||||||
# MoveIt motion planning
|
|
||||||
- ros-humble-moveit
|
|
||||||
- ros-humble-moveit-servo
|
|
||||||
# Simulation
|
|
||||||
- ros-humble-simulation
|
|
||||||
|
|
||||||
about:
|
|
||||||
repository: https://github.com/deepmodeling/Uni-Lab-OS
|
|
||||||
license: GPL-3.0-only
|
|
||||||
description: "UniLabOS Full - Complete package with ROS2 Desktop, MoveIt, Navigation2, Gazebo, Jupyter"
|
|
||||||
91
.conda/recipe.yaml
Normal file
91
.conda/recipe.yaml
Normal file
@@ -0,0 +1,91 @@
|
|||||||
|
package:
|
||||||
|
name: unilabos
|
||||||
|
version: 0.10.15
|
||||||
|
|
||||||
|
source:
|
||||||
|
path: ../unilabos
|
||||||
|
target_directory: unilabos
|
||||||
|
|
||||||
|
build:
|
||||||
|
python:
|
||||||
|
entry_points:
|
||||||
|
- unilab = unilabos.app.main:main
|
||||||
|
script:
|
||||||
|
- set PIP_NO_INDEX=
|
||||||
|
- if: win
|
||||||
|
then:
|
||||||
|
- copy %RECIPE_DIR%\..\MANIFEST.in %SRC_DIR%
|
||||||
|
- copy %RECIPE_DIR%\..\setup.cfg %SRC_DIR%
|
||||||
|
- copy %RECIPE_DIR%\..\setup.py %SRC_DIR%
|
||||||
|
- call %PYTHON% -m pip install %SRC_DIR%
|
||||||
|
- if: unix
|
||||||
|
then:
|
||||||
|
- cp $RECIPE_DIR/../MANIFEST.in $SRC_DIR
|
||||||
|
- cp $RECIPE_DIR/../setup.cfg $SRC_DIR
|
||||||
|
- cp $RECIPE_DIR/../setup.py $SRC_DIR
|
||||||
|
- $PYTHON -m pip install $SRC_DIR
|
||||||
|
|
||||||
|
requirements:
|
||||||
|
host:
|
||||||
|
- python ==3.11.11
|
||||||
|
- pip
|
||||||
|
- setuptools
|
||||||
|
- zstd
|
||||||
|
- zstandard
|
||||||
|
run:
|
||||||
|
- conda-forge::python ==3.11.11
|
||||||
|
- compilers
|
||||||
|
- cmake
|
||||||
|
- zstd
|
||||||
|
- zstandard
|
||||||
|
- ninja
|
||||||
|
- if: unix
|
||||||
|
then:
|
||||||
|
- make
|
||||||
|
- sphinx
|
||||||
|
- sphinx_rtd_theme
|
||||||
|
- numpy
|
||||||
|
- scipy
|
||||||
|
- pandas
|
||||||
|
- networkx
|
||||||
|
- matplotlib
|
||||||
|
- pint
|
||||||
|
- pyserial
|
||||||
|
- pyusb
|
||||||
|
- pylibftdi
|
||||||
|
- pymodbus
|
||||||
|
- python-can
|
||||||
|
- pyvisa
|
||||||
|
- opencv
|
||||||
|
- pydantic
|
||||||
|
- fastapi
|
||||||
|
- uvicorn
|
||||||
|
- gradio
|
||||||
|
- flask
|
||||||
|
- websockets
|
||||||
|
- ipython
|
||||||
|
- jupyter
|
||||||
|
- jupyros
|
||||||
|
- colcon-common-extensions
|
||||||
|
- robostack-staging::ros-humble-desktop-full
|
||||||
|
- robostack-staging::ros-humble-control-msgs
|
||||||
|
- robostack-staging::ros-humble-sensor-msgs
|
||||||
|
- robostack-staging::ros-humble-trajectory-msgs
|
||||||
|
- ros-humble-navigation2
|
||||||
|
- ros-humble-ros2-control
|
||||||
|
- ros-humble-robot-state-publisher
|
||||||
|
- ros-humble-joint-state-publisher
|
||||||
|
- ros-humble-rosbridge-server
|
||||||
|
- ros-humble-cv-bridge
|
||||||
|
- ros-humble-tf2
|
||||||
|
- ros-humble-moveit
|
||||||
|
- ros-humble-moveit-servo
|
||||||
|
- ros-humble-simulation
|
||||||
|
- ros-humble-tf-transformations
|
||||||
|
- transforms3d
|
||||||
|
- uni-lab::ros-humble-unilabos-msgs
|
||||||
|
|
||||||
|
about:
|
||||||
|
repository: https://github.com/deepmodeling/Uni-Lab-OS
|
||||||
|
license: GPL-3.0-only
|
||||||
|
description: "Uni-Lab-OS"
|
||||||
39
.github/workflows/ci-check.yml
vendored
39
.github/workflows/ci-check.yml
vendored
@@ -8,19 +8,14 @@ on:
|
|||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
registry-check:
|
registry-check:
|
||||||
runs-on: windows-latest
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
env:
|
|
||||||
# Fix Unicode encoding issue on Windows runner (cp1252 -> utf-8)
|
|
||||||
PYTHONIOENCODING: utf-8
|
|
||||||
PYTHONUTF8: 1
|
|
||||||
|
|
||||||
defaults:
|
defaults:
|
||||||
run:
|
run:
|
||||||
shell: cmd
|
shell: bash -l {0}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
@@ -32,31 +27,33 @@ jobs:
|
|||||||
channels: robostack-staging,conda-forge,uni-lab
|
channels: robostack-staging,conda-forge,uni-lab
|
||||||
channel-priority: flexible
|
channel-priority: flexible
|
||||||
activate-environment: check-env
|
activate-environment: check-env
|
||||||
|
auto-activate-base: false
|
||||||
auto-update-conda: false
|
auto-update-conda: false
|
||||||
show-channel-urls: true
|
show-channel-urls: true
|
||||||
|
|
||||||
- name: Install ROS dependencies, uv and unilabos-msgs
|
- name: Install ROS dependencies and unilabos-msgs
|
||||||
run: |
|
run: |
|
||||||
echo Installing ROS dependencies...
|
# Install all packages together for proper dependency resolution
|
||||||
mamba install -n check-env conda-forge::uv conda-forge::opencv robostack-staging::ros-humble-ros-core robostack-staging::ros-humble-action-msgs robostack-staging::ros-humble-std-msgs robostack-staging::ros-humble-geometry-msgs robostack-staging::ros-humble-control-msgs robostack-staging::ros-humble-nav2-msgs uni-lab::ros-humble-unilabos-msgs robostack-staging::ros-humble-cv-bridge robostack-staging::ros-humble-vision-opencv robostack-staging::ros-humble-tf-transformations robostack-staging::ros-humble-moveit-msgs robostack-staging::ros-humble-tf2-ros robostack-staging::ros-humble-tf2-ros-py conda-forge::transforms3d -c robostack-staging -c conda-forge -c uni-lab -y
|
# Use mamba for faster and more reliable solving
|
||||||
|
mamba install -n check-env \
|
||||||
|
python=3.11.11 \
|
||||||
|
robostack-staging::ros-humble-ros-core \
|
||||||
|
robostack-staging::ros-humble-action-msgs \
|
||||||
|
robostack-staging::ros-humble-std-msgs \
|
||||||
|
robostack-staging::ros-humble-geometry-msgs \
|
||||||
|
robostack-staging::ros-humble-control-msgs \
|
||||||
|
uni-lab::ros-humble-unilabos-msgs \
|
||||||
|
-c robostack-staging -c conda-forge -c uni-lab -y
|
||||||
|
|
||||||
- name: Install pip dependencies and unilabos
|
- name: Install unilabos project
|
||||||
run: |
|
run: |
|
||||||
call conda activate check-env
|
pip install -e .
|
||||||
echo Installing pip dependencies...
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
uv pip install pywinauto git+https://github.com/Xuwznln/pylabrobot.git
|
|
||||||
uv pip uninstall enum34 || echo enum34 not installed, skipping
|
|
||||||
uv pip install .
|
|
||||||
|
|
||||||
- name: Run check mode (complete_registry)
|
- name: Run check mode (complete_registry)
|
||||||
run: |
|
run: |
|
||||||
call conda activate check-env
|
|
||||||
echo Running check mode...
|
|
||||||
python -m unilabos --check_mode --skip_env_check
|
python -m unilabos --check_mode --skip_env_check
|
||||||
|
|
||||||
- name: Check for uncommitted changes
|
- name: Check for uncommitted changes
|
||||||
shell: bash
|
|
||||||
run: |
|
run: |
|
||||||
if ! git diff --exit-code; then
|
if ! git diff --exit-code; then
|
||||||
echo "::error::检测到文件变化!请先在本地运行 'python -m unilabos --complete_registry' 并提交变更"
|
echo "::error::检测到文件变化!请先在本地运行 'python -m unilabos --complete_registry' 并提交变更"
|
||||||
|
|||||||
39
.github/workflows/conda-pack-build.yml
vendored
39
.github/workflows/conda-pack-build.yml
vendored
@@ -13,11 +13,6 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: 'win-64'
|
default: 'win-64'
|
||||||
type: string
|
type: string
|
||||||
build_full:
|
|
||||||
description: '是否构建完整版 unilabos-full (默认构建轻量版 unilabos)'
|
|
||||||
required: false
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
build-conda-pack:
|
build-conda-pack:
|
||||||
@@ -62,7 +57,7 @@ jobs:
|
|||||||
echo "should_build=false" >> $GITHUB_OUTPUT
|
echo "should_build=false" >> $GITHUB_OUTPUT
|
||||||
fi
|
fi
|
||||||
|
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
with:
|
with:
|
||||||
ref: ${{ github.event.inputs.branch }}
|
ref: ${{ github.event.inputs.branch }}
|
||||||
@@ -74,7 +69,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
miniforge-version: latest
|
miniforge-version: latest
|
||||||
use-mamba: true
|
use-mamba: true
|
||||||
python-version: '3.11.14'
|
python-version: '3.11.11'
|
||||||
channels: conda-forge,robostack-staging,uni-lab,defaults
|
channels: conda-forge,robostack-staging,uni-lab,defaults
|
||||||
channel-priority: flexible
|
channel-priority: flexible
|
||||||
activate-environment: unilab
|
activate-environment: unilab
|
||||||
@@ -86,14 +81,7 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
echo Installing unilabos and dependencies to unilab environment...
|
echo Installing unilabos and dependencies to unilab environment...
|
||||||
echo Using mamba for faster and more reliable dependency resolution...
|
echo Using mamba for faster and more reliable dependency resolution...
|
||||||
echo Build full: ${{ github.event.inputs.build_full }}
|
|
||||||
if "${{ github.event.inputs.build_full }}"=="true" (
|
|
||||||
echo Installing unilabos-full ^(complete package^)...
|
|
||||||
mamba install -n unilab uni-lab::unilabos-full conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
|
||||||
) else (
|
|
||||||
echo Installing unilabos ^(minimal package^)...
|
|
||||||
mamba install -n unilab uni-lab::unilabos conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
mamba install -n unilab uni-lab::unilabos conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
||||||
)
|
|
||||||
|
|
||||||
- name: Install conda-pack, unilabos and dependencies (Unix)
|
- name: Install conda-pack, unilabos and dependencies (Unix)
|
||||||
if: steps.should_build.outputs.should_build == 'true' && matrix.platform != 'win-64'
|
if: steps.should_build.outputs.should_build == 'true' && matrix.platform != 'win-64'
|
||||||
@@ -101,14 +89,7 @@ jobs:
|
|||||||
run: |
|
run: |
|
||||||
echo "Installing unilabos and dependencies to unilab environment..."
|
echo "Installing unilabos and dependencies to unilab environment..."
|
||||||
echo "Using mamba for faster and more reliable dependency resolution..."
|
echo "Using mamba for faster and more reliable dependency resolution..."
|
||||||
echo "Build full: ${{ github.event.inputs.build_full }}"
|
|
||||||
if [[ "${{ github.event.inputs.build_full }}" == "true" ]]; then
|
|
||||||
echo "Installing unilabos-full (complete package)..."
|
|
||||||
mamba install -n unilab uni-lab::unilabos-full conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
|
||||||
else
|
|
||||||
echo "Installing unilabos (minimal package)..."
|
|
||||||
mamba install -n unilab uni-lab::unilabos conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
mamba install -n unilab uni-lab::unilabos conda-pack -c uni-lab -c robostack-staging -c conda-forge -y
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Get latest ros-humble-unilabos-msgs version (Windows)
|
- name: Get latest ros-humble-unilabos-msgs version (Windows)
|
||||||
if: steps.should_build.outputs.should_build == 'true' && matrix.platform == 'win-64'
|
if: steps.should_build.outputs.should_build == 'true' && matrix.platform == 'win-64'
|
||||||
@@ -312,7 +293,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Upload distribution package
|
- name: Upload distribution package
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
uses: actions/upload-artifact@v6
|
uses: actions/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: unilab-pack-${{ matrix.platform }}-${{ github.event.inputs.branch }}
|
name: unilab-pack-${{ matrix.platform }}-${{ github.event.inputs.branch }}
|
||||||
path: dist-package/
|
path: dist-package/
|
||||||
@@ -327,12 +308,7 @@ jobs:
|
|||||||
echo ==========================================
|
echo ==========================================
|
||||||
echo Platform: ${{ matrix.platform }}
|
echo Platform: ${{ matrix.platform }}
|
||||||
echo Branch: ${{ github.event.inputs.branch }}
|
echo Branch: ${{ github.event.inputs.branch }}
|
||||||
echo Python version: 3.11.14
|
echo Python version: 3.11.11
|
||||||
if "${{ github.event.inputs.build_full }}"=="true" (
|
|
||||||
echo Package: unilabos-full ^(complete^)
|
|
||||||
) else (
|
|
||||||
echo Package: unilabos ^(minimal^)
|
|
||||||
)
|
|
||||||
echo.
|
echo.
|
||||||
echo Distribution package contents:
|
echo Distribution package contents:
|
||||||
dir dist-package
|
dir dist-package
|
||||||
@@ -352,12 +328,7 @@ jobs:
|
|||||||
echo "=========================================="
|
echo "=========================================="
|
||||||
echo "Platform: ${{ matrix.platform }}"
|
echo "Platform: ${{ matrix.platform }}"
|
||||||
echo "Branch: ${{ github.event.inputs.branch }}"
|
echo "Branch: ${{ github.event.inputs.branch }}"
|
||||||
echo "Python version: 3.11.14"
|
echo "Python version: 3.11.11"
|
||||||
if [[ "${{ github.event.inputs.build_full }}" == "true" ]]; then
|
|
||||||
echo "Package: unilabos-full (complete)"
|
|
||||||
else
|
|
||||||
echo "Package: unilabos (minimal)"
|
|
||||||
fi
|
|
||||||
echo ""
|
echo ""
|
||||||
echo "Distribution package contents:"
|
echo "Distribution package contents:"
|
||||||
ls -lh dist-package/
|
ls -lh dist-package/
|
||||||
|
|||||||
37
.github/workflows/deploy-docs.yml
vendored
37
.github/workflows/deploy-docs.yml
vendored
@@ -1,12 +1,10 @@
|
|||||||
name: Deploy Docs
|
name: Deploy Docs
|
||||||
|
|
||||||
on:
|
on:
|
||||||
# 在 CI Check 成功后自动触发(仅 main 分支)
|
push:
|
||||||
workflow_run:
|
branches: [main]
|
||||||
workflows: ["CI Check"]
|
pull_request:
|
||||||
types: [completed]
|
|
||||||
branches: [main]
|
branches: [main]
|
||||||
# 手动触发
|
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
branch:
|
branch:
|
||||||
@@ -35,19 +33,12 @@ concurrency:
|
|||||||
jobs:
|
jobs:
|
||||||
# Build documentation
|
# Build documentation
|
||||||
build:
|
build:
|
||||||
# 只在以下情况运行:
|
|
||||||
# 1. workflow_run 触发且 CI Check 成功
|
|
||||||
# 2. 手动触发
|
|
||||||
if: |
|
|
||||||
github.event_name == 'workflow_dispatch' ||
|
|
||||||
(github.event_name == 'workflow_run' && github.event.workflow_run.conclusion == 'success')
|
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout code
|
- name: Checkout code
|
||||||
uses: actions/checkout@v6
|
uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
# workflow_run 时使用触发工作流的分支,手动触发时使用输入的分支
|
ref: ${{ github.event.inputs.branch || github.ref }}
|
||||||
ref: ${{ github.event.workflow_run.head_branch || github.event.inputs.branch || github.ref }}
|
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
- name: Setup Miniforge (with mamba)
|
- name: Setup Miniforge (with mamba)
|
||||||
@@ -55,7 +46,7 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
miniforge-version: latest
|
miniforge-version: latest
|
||||||
use-mamba: true
|
use-mamba: true
|
||||||
python-version: '3.11.14'
|
python-version: '3.11.11'
|
||||||
channels: conda-forge,robostack-staging,uni-lab,defaults
|
channels: conda-forge,robostack-staging,uni-lab,defaults
|
||||||
channel-priority: flexible
|
channel-priority: flexible
|
||||||
activate-environment: unilab
|
activate-environment: unilab
|
||||||
@@ -84,10 +75,8 @@ jobs:
|
|||||||
|
|
||||||
- name: Setup Pages
|
- name: Setup Pages
|
||||||
id: pages
|
id: pages
|
||||||
uses: actions/configure-pages@v5
|
uses: actions/configure-pages@v4
|
||||||
if: |
|
if: github.ref == 'refs/heads/main' || (github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
||||||
github.event.workflow_run.head_branch == 'main' ||
|
|
||||||
(github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
|
||||||
|
|
||||||
- name: Build Sphinx documentation
|
- name: Build Sphinx documentation
|
||||||
run: |
|
run: |
|
||||||
@@ -105,18 +94,14 @@ jobs:
|
|||||||
test -f docs/_build/html/index.html && echo "✓ index.html exists" || echo "✗ index.html missing"
|
test -f docs/_build/html/index.html && echo "✓ index.html exists" || echo "✗ index.html missing"
|
||||||
|
|
||||||
- name: Upload build artifacts
|
- name: Upload build artifacts
|
||||||
uses: actions/upload-pages-artifact@v4
|
uses: actions/upload-pages-artifact@v3
|
||||||
if: |
|
if: github.ref == 'refs/heads/main' || (github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
||||||
github.event.workflow_run.head_branch == 'main' ||
|
|
||||||
(github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
|
||||||
with:
|
with:
|
||||||
path: docs/_build/html
|
path: docs/_build/html
|
||||||
|
|
||||||
# Deploy to GitHub Pages
|
# Deploy to GitHub Pages
|
||||||
deploy:
|
deploy:
|
||||||
if: |
|
if: github.ref == 'refs/heads/main' || (github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
||||||
github.event.workflow_run.head_branch == 'main' ||
|
|
||||||
(github.event_name == 'workflow_dispatch' && github.event.inputs.deploy_to_pages == 'true')
|
|
||||||
environment:
|
environment:
|
||||||
name: github-pages
|
name: github-pages
|
||||||
url: ${{ steps.deployment.outputs.page_url }}
|
url: ${{ steps.deployment.outputs.page_url }}
|
||||||
|
|||||||
46
.github/workflows/multi-platform-build.yml
vendored
46
.github/workflows/multi-platform-build.yml
vendored
@@ -1,16 +1,11 @@
|
|||||||
name: Multi-Platform Conda Build
|
name: Multi-Platform Conda Build
|
||||||
|
|
||||||
on:
|
on:
|
||||||
# 在 CI Check 工作流完成后触发(仅限 main/dev 分支)
|
|
||||||
workflow_run:
|
|
||||||
workflows: ["CI Check"]
|
|
||||||
types:
|
|
||||||
- completed
|
|
||||||
branches: [main, dev]
|
|
||||||
# 支持 tag 推送(不依赖 CI Check)
|
|
||||||
push:
|
push:
|
||||||
|
branches: [main, dev]
|
||||||
tags: ['v*']
|
tags: ['v*']
|
||||||
# 手动触发
|
pull_request:
|
||||||
|
branches: [main, dev]
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
platforms:
|
platforms:
|
||||||
@@ -22,37 +17,9 @@ on:
|
|||||||
required: false
|
required: false
|
||||||
default: false
|
default: false
|
||||||
type: boolean
|
type: boolean
|
||||||
skip_ci_check:
|
|
||||||
description: '跳过等待 CI Check (手动触发时可选)'
|
|
||||||
required: false
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
# 等待 CI Check 完成的 job (仅用于 workflow_run 触发)
|
|
||||||
wait-for-ci:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
if: github.event_name == 'workflow_run'
|
|
||||||
outputs:
|
|
||||||
should_continue: ${{ steps.check.outputs.should_continue }}
|
|
||||||
steps:
|
|
||||||
- name: Check CI status
|
|
||||||
id: check
|
|
||||||
run: |
|
|
||||||
if [[ "${{ github.event.workflow_run.conclusion }}" == "success" ]]; then
|
|
||||||
echo "should_continue=true" >> $GITHUB_OUTPUT
|
|
||||||
echo "CI Check passed, proceeding with build"
|
|
||||||
else
|
|
||||||
echo "should_continue=false" >> $GITHUB_OUTPUT
|
|
||||||
echo "CI Check did not succeed (status: ${{ github.event.workflow_run.conclusion }}), skipping build"
|
|
||||||
fi
|
|
||||||
|
|
||||||
build:
|
build:
|
||||||
needs: [wait-for-ci]
|
|
||||||
# 运行条件:workflow_run 触发且 CI 成功,或者其他触发方式
|
|
||||||
if: |
|
|
||||||
always() &&
|
|
||||||
(needs.wait-for-ci.result == 'skipped' || needs.wait-for-ci.outputs.should_continue == 'true')
|
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@@ -77,10 +44,8 @@ jobs:
|
|||||||
shell: bash -l {0}
|
shell: bash -l {0}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
# 如果是 workflow_run 触发,使用触发 CI Check 的 commit
|
|
||||||
ref: ${{ github.event.workflow_run.head_sha || github.ref }}
|
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
- name: Check if platform should be built
|
- name: Check if platform should be built
|
||||||
@@ -104,6 +69,7 @@ jobs:
|
|||||||
channels: conda-forge,robostack-staging,defaults
|
channels: conda-forge,robostack-staging,defaults
|
||||||
channel-priority: strict
|
channel-priority: strict
|
||||||
activate-environment: build-env
|
activate-environment: build-env
|
||||||
|
auto-activate-base: false
|
||||||
auto-update-conda: false
|
auto-update-conda: false
|
||||||
show-channel-urls: true
|
show-channel-urls: true
|
||||||
|
|
||||||
@@ -149,7 +115,7 @@ jobs:
|
|||||||
|
|
||||||
- name: Upload conda package artifacts
|
- name: Upload conda package artifacts
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
uses: actions/upload-artifact@v6
|
uses: actions/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: conda-package-${{ matrix.platform }}
|
name: conda-package-${{ matrix.platform }}
|
||||||
path: conda-packages-temp
|
path: conda-packages-temp
|
||||||
|
|||||||
113
.github/workflows/unilabos-conda-build.yml
vendored
113
.github/workflows/unilabos-conda-build.yml
vendored
@@ -1,62 +1,25 @@
|
|||||||
name: UniLabOS Conda Build
|
name: UniLabOS Conda Build
|
||||||
|
|
||||||
on:
|
on:
|
||||||
# 在 CI Check 成功后自动触发
|
|
||||||
workflow_run:
|
|
||||||
workflows: ["CI Check"]
|
|
||||||
types: [completed]
|
|
||||||
branches: [main, dev]
|
|
||||||
# 标签推送时直接触发(发布版本)
|
|
||||||
push:
|
push:
|
||||||
|
branches: [main, dev]
|
||||||
tags: ['v*']
|
tags: ['v*']
|
||||||
# 手动触发
|
pull_request:
|
||||||
|
branches: [main, dev]
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
platforms:
|
platforms:
|
||||||
description: '选择构建平台 (逗号分隔): linux-64, osx-64, osx-arm64, win-64'
|
description: '选择构建平台 (逗号分隔): linux-64, osx-64, osx-arm64, win-64'
|
||||||
required: false
|
required: false
|
||||||
default: 'linux-64'
|
default: 'linux-64'
|
||||||
build_full:
|
|
||||||
description: '是否构建 unilabos-full 完整包 (默认只构建 unilabos 基础包)'
|
|
||||||
required: false
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
upload_to_anaconda:
|
upload_to_anaconda:
|
||||||
description: '是否上传到Anaconda.org'
|
description: '是否上传到Anaconda.org'
|
||||||
required: false
|
required: false
|
||||||
default: false
|
default: false
|
||||||
type: boolean
|
type: boolean
|
||||||
skip_ci_check:
|
|
||||||
description: '跳过等待 CI Check (手动触发时可选)'
|
|
||||||
required: false
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
# 等待 CI Check 完成的 job (仅用于 workflow_run 触发)
|
|
||||||
wait-for-ci:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
if: github.event_name == 'workflow_run'
|
|
||||||
outputs:
|
|
||||||
should_continue: ${{ steps.check.outputs.should_continue }}
|
|
||||||
steps:
|
|
||||||
- name: Check CI status
|
|
||||||
id: check
|
|
||||||
run: |
|
|
||||||
if [[ "${{ github.event.workflow_run.conclusion }}" == "success" ]]; then
|
|
||||||
echo "should_continue=true" >> $GITHUB_OUTPUT
|
|
||||||
echo "CI Check passed, proceeding with build"
|
|
||||||
else
|
|
||||||
echo "should_continue=false" >> $GITHUB_OUTPUT
|
|
||||||
echo "CI Check did not succeed (status: ${{ github.event.workflow_run.conclusion }}), skipping build"
|
|
||||||
fi
|
|
||||||
|
|
||||||
build:
|
build:
|
||||||
needs: [wait-for-ci]
|
|
||||||
# 运行条件:workflow_run 触发且 CI 成功,或者其他触发方式
|
|
||||||
if: |
|
|
||||||
always() &&
|
|
||||||
(needs.wait-for-ci.result == 'skipped' || needs.wait-for-ci.outputs.should_continue == 'true')
|
|
||||||
strategy:
|
strategy:
|
||||||
fail-fast: false
|
fail-fast: false
|
||||||
matrix:
|
matrix:
|
||||||
@@ -77,10 +40,8 @@ jobs:
|
|||||||
shell: bash -l {0}
|
shell: bash -l {0}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v6
|
- uses: actions/checkout@v4
|
||||||
with:
|
with:
|
||||||
# 如果是 workflow_run 触发,使用触发 CI Check 的 commit
|
|
||||||
ref: ${{ github.event.workflow_run.head_sha || github.ref }}
|
|
||||||
fetch-depth: 0
|
fetch-depth: 0
|
||||||
|
|
||||||
- name: Check if platform should be built
|
- name: Check if platform should be built
|
||||||
@@ -104,6 +65,7 @@ jobs:
|
|||||||
channels: conda-forge,robostack-staging,uni-lab,defaults
|
channels: conda-forge,robostack-staging,uni-lab,defaults
|
||||||
channel-priority: strict
|
channel-priority: strict
|
||||||
activate-environment: build-env
|
activate-environment: build-env
|
||||||
|
auto-activate-base: false
|
||||||
auto-update-conda: false
|
auto-update-conda: false
|
||||||
show-channel-urls: true
|
show-channel-urls: true
|
||||||
|
|
||||||
@@ -119,61 +81,12 @@ jobs:
|
|||||||
conda list | grep -E "(rattler-build|anaconda-client)"
|
conda list | grep -E "(rattler-build|anaconda-client)"
|
||||||
echo "Platform: ${{ matrix.platform }}"
|
echo "Platform: ${{ matrix.platform }}"
|
||||||
echo "OS: ${{ matrix.os }}"
|
echo "OS: ${{ matrix.os }}"
|
||||||
echo "Build full package: ${{ github.event.inputs.build_full || 'false' }}"
|
echo "Building UniLabOS package"
|
||||||
echo "Building packages:"
|
|
||||||
echo " - unilabos-env (environment dependencies)"
|
|
||||||
echo " - unilabos (with pip package)"
|
|
||||||
if [[ "${{ github.event.inputs.build_full }}" == "true" ]]; then
|
|
||||||
echo " - unilabos-full (complete package)"
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Build unilabos-env (conda environment only, noarch)
|
- name: Build conda package
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
run: |
|
run: |
|
||||||
echo "Building unilabos-env (conda environment dependencies)..."
|
rattler-build build -r .conda/recipe.yaml -c uni-lab -c robostack-staging -c conda-forge
|
||||||
rattler-build build -r .conda/environment/recipe.yaml -c uni-lab -c robostack-staging -c conda-forge
|
|
||||||
|
|
||||||
- name: Upload unilabos-env to Anaconda.org (if enabled)
|
|
||||||
if: steps.should_build.outputs.should_build == 'true' && github.event.inputs.upload_to_anaconda == 'true'
|
|
||||||
run: |
|
|
||||||
echo "Uploading unilabos-env to uni-lab organization..."
|
|
||||||
for package in $(find ./output -name "unilabos-env*.conda"); do
|
|
||||||
anaconda -t ${{ secrets.ANACONDA_API_TOKEN }} upload --user uni-lab --force "$package"
|
|
||||||
done
|
|
||||||
|
|
||||||
- name: Build unilabos (with pip package)
|
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
|
||||||
run: |
|
|
||||||
echo "Building unilabos package..."
|
|
||||||
# 如果已上传到 Anaconda,从 uni-lab channel 获取 unilabos-env;否则从本地 output 获取
|
|
||||||
rattler-build build -r .conda/base/recipe.yaml -c uni-lab -c robostack-staging -c conda-forge --channel ./output
|
|
||||||
|
|
||||||
- name: Upload unilabos to Anaconda.org (if enabled)
|
|
||||||
if: steps.should_build.outputs.should_build == 'true' && github.event.inputs.upload_to_anaconda == 'true'
|
|
||||||
run: |
|
|
||||||
echo "Uploading unilabos to uni-lab organization..."
|
|
||||||
for package in $(find ./output -name "unilabos-0*.conda" -o -name "unilabos-[0-9]*.conda"); do
|
|
||||||
anaconda -t ${{ secrets.ANACONDA_API_TOKEN }} upload --user uni-lab --force "$package"
|
|
||||||
done
|
|
||||||
|
|
||||||
- name: Build unilabos-full - Only when explicitly requested
|
|
||||||
if: |
|
|
||||||
steps.should_build.outputs.should_build == 'true' &&
|
|
||||||
github.event.inputs.build_full == 'true'
|
|
||||||
run: |
|
|
||||||
echo "Building unilabos-full package on ${{ matrix.platform }}..."
|
|
||||||
rattler-build build -r .conda/full/recipe.yaml -c uni-lab -c robostack-staging -c conda-forge --channel ./output
|
|
||||||
|
|
||||||
- name: Upload unilabos-full to Anaconda.org (if enabled)
|
|
||||||
if: |
|
|
||||||
steps.should_build.outputs.should_build == 'true' &&
|
|
||||||
github.event.inputs.build_full == 'true' &&
|
|
||||||
github.event.inputs.upload_to_anaconda == 'true'
|
|
||||||
run: |
|
|
||||||
echo "Uploading unilabos-full to uni-lab organization..."
|
|
||||||
for package in $(find ./output -name "unilabos-full*.conda"); do
|
|
||||||
anaconda -t ${{ secrets.ANACONDA_API_TOKEN }} upload --user uni-lab --force "$package"
|
|
||||||
done
|
|
||||||
|
|
||||||
- name: List built packages
|
- name: List built packages
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
@@ -195,9 +108,17 @@ jobs:
|
|||||||
|
|
||||||
- name: Upload conda package artifacts
|
- name: Upload conda package artifacts
|
||||||
if: steps.should_build.outputs.should_build == 'true'
|
if: steps.should_build.outputs.should_build == 'true'
|
||||||
uses: actions/upload-artifact@v6
|
uses: actions/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: conda-package-unilabos-${{ matrix.platform }}
|
name: conda-package-unilabos-${{ matrix.platform }}
|
||||||
path: conda-packages-temp
|
path: conda-packages-temp
|
||||||
if-no-files-found: warn
|
if-no-files-found: warn
|
||||||
retention-days: 30
|
retention-days: 30
|
||||||
|
|
||||||
|
- name: Upload to Anaconda.org (uni-lab organization)
|
||||||
|
if: github.event.inputs.upload_to_anaconda == 'true'
|
||||||
|
run: |
|
||||||
|
for package in $(find ./output -name "*.conda"); do
|
||||||
|
echo "Uploading $package to uni-lab organization..."
|
||||||
|
anaconda -t ${{ secrets.ANACONDA_API_TOKEN }} upload --user uni-lab --force "$package"
|
||||||
|
done
|
||||||
|
|||||||
@@ -1,5 +1,4 @@
|
|||||||
recursive-include unilabos/test *
|
recursive-include unilabos/test *
|
||||||
recursive-include unilabos/utils *
|
|
||||||
recursive-include unilabos/registry *.yaml
|
recursive-include unilabos/registry *.yaml
|
||||||
recursive-include unilabos/app/web/static *
|
recursive-include unilabos/app/web/static *
|
||||||
recursive-include unilabos/app/web/templates *
|
recursive-include unilabos/app/web/templates *
|
||||||
|
|||||||
38
README.md
38
README.md
@@ -31,46 +31,26 @@ Detailed documentation can be found at:
|
|||||||
|
|
||||||
## Quick Start
|
## Quick Start
|
||||||
|
|
||||||
### 1. Setup Conda Environment
|
1. Setup Conda Environment
|
||||||
|
|
||||||
Uni-Lab-OS recommends using `mamba` for environment management. Choose the package that fits your needs:
|
Uni-Lab-OS recommends using `mamba` for environment management:
|
||||||
|
|
||||||
| Package | Use Case | Contents |
|
|
||||||
|---------|----------|----------|
|
|
||||||
| `unilabos` | **Recommended for most users** | Complete package, ready to use |
|
|
||||||
| `unilabos-env` | Developers (editable install) | Environment only, install unilabos via pip |
|
|
||||||
| `unilabos-full` | Simulation/Visualization | unilabos + ROS2 Desktop + Gazebo + MoveIt |
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# Create new environment
|
# Create new environment
|
||||||
mamba create -n unilab python=3.11.14
|
mamba create -n unilab python=3.11.11
|
||||||
mamba activate unilab
|
mamba activate unilab
|
||||||
|
mamba install -n unilab uni-lab::unilabos -c robostack-staging -c conda-forge
|
||||||
# Option A: Standard installation (recommended for most users)
|
|
||||||
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
|
||||||
|
|
||||||
# Option B: For developers (editable mode development)
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
# Then install unilabos and dependencies:
|
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git && cd Uni-Lab-OS
|
|
||||||
pip install -e .
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
|
|
||||||
# Option C: Full installation (simulation/visualization)
|
|
||||||
mamba install uni-lab::unilabos-full -c robostack-staging -c conda-forge
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**When to use which?**
|
2. Install Dev Uni-Lab-OS
|
||||||
- **unilabos**: Standard installation for production deployment and general usage (recommended)
|
|
||||||
- **unilabos-env**: For developers who need `pip install -e .` editable mode, modify source code
|
|
||||||
- **unilabos-full**: For simulation (Gazebo), visualization (rviz2), and Jupyter notebooks
|
|
||||||
|
|
||||||
### 2. Clone Repository (Optional, for developers)
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# Clone the repository (only needed for development or examples)
|
# Clone the repository
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
||||||
cd Uni-Lab-OS
|
cd Uni-Lab-OS
|
||||||
|
|
||||||
|
# Install Uni-Lab-OS
|
||||||
|
pip install .
|
||||||
```
|
```
|
||||||
|
|
||||||
3. Start Uni-Lab System
|
3. Start Uni-Lab System
|
||||||
|
|||||||
38
README_zh.md
38
README_zh.md
@@ -31,46 +31,26 @@ Uni-Lab-OS 是一个用于实验室自动化的综合平台,旨在连接和控
|
|||||||
|
|
||||||
## 快速开始
|
## 快速开始
|
||||||
|
|
||||||
### 1. 配置 Conda 环境
|
1. 配置 Conda 环境
|
||||||
|
|
||||||
Uni-Lab-OS 建议使用 `mamba` 管理环境。根据您的需求选择合适的安装包:
|
Uni-Lab-OS 建议使用 `mamba` 管理环境。根据您的操作系统选择适当的环境文件:
|
||||||
|
|
||||||
| 安装包 | 适用场景 | 包含内容 |
|
|
||||||
|--------|----------|----------|
|
|
||||||
| `unilabos` | **推荐大多数用户** | 完整安装包,开箱即用 |
|
|
||||||
| `unilabos-env` | 开发者(可编辑安装) | 仅环境依赖,通过 pip 安装 unilabos |
|
|
||||||
| `unilabos-full` | 仿真/可视化 | unilabos + ROS2 桌面版 + Gazebo + MoveIt |
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# 创建新环境
|
# 创建新环境
|
||||||
mamba create -n unilab python=3.11.14
|
mamba create -n unilab python=3.11.11
|
||||||
mamba activate unilab
|
mamba activate unilab
|
||||||
|
mamba install -n unilab uni-lab::unilabos -c robostack-staging -c conda-forge
|
||||||
# 方案 A:标准安装(推荐大多数用户)
|
|
||||||
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
|
||||||
|
|
||||||
# 方案 B:开发者环境(可编辑模式开发)
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
# 然后安装 unilabos 和依赖:
|
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git && cd Uni-Lab-OS
|
|
||||||
pip install -e .
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
|
|
||||||
# 方案 C:完整安装(仿真/可视化)
|
|
||||||
mamba install uni-lab::unilabos-full -c robostack-staging -c conda-forge
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**如何选择?**
|
2. 安装开发版 Uni-Lab-OS:
|
||||||
- **unilabos**:标准安装,适用于生产部署和日常使用(推荐)
|
|
||||||
- **unilabos-env**:开发者使用,支持 `pip install -e .` 可编辑模式,可修改源代码
|
|
||||||
- **unilabos-full**:需要仿真(Gazebo)、可视化(rviz2)或 Jupyter Notebook
|
|
||||||
|
|
||||||
### 2. 克隆仓库(可选,供开发者使用)
|
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# 克隆仓库(仅开发或查看示例时需要)
|
# 克隆仓库
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
||||||
cd Uni-Lab-OS
|
cd Uni-Lab-OS
|
||||||
|
|
||||||
|
# 安装 Uni-Lab-OS
|
||||||
|
pip install .
|
||||||
```
|
```
|
||||||
|
|
||||||
3. 启动 Uni-Lab 系统
|
3. 启动 Uni-Lab 系统
|
||||||
|
|||||||
@@ -31,14 +31,6 @@
|
|||||||
|
|
||||||
详细的安装步骤请参考 [安装指南](installation.md)。
|
详细的安装步骤请参考 [安装指南](installation.md)。
|
||||||
|
|
||||||
**选择合适的安装包:**
|
|
||||||
|
|
||||||
| 安装包 | 适用场景 | 包含组件 |
|
|
||||||
|--------|----------|----------|
|
|
||||||
| `unilabos` | **推荐大多数用户**,生产部署 | 完整安装包,开箱即用 |
|
|
||||||
| `unilabos-env` | 开发者(可编辑安装) | 仅环境依赖,通过 pip 安装 unilabos |
|
|
||||||
| `unilabos-full` | 仿真/可视化 | unilabos + 完整 ROS2 桌面版 + Gazebo + MoveIt |
|
|
||||||
|
|
||||||
**关键步骤:**
|
**关键步骤:**
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -46,30 +38,15 @@
|
|||||||
# 下载 Miniforge: https://github.com/conda-forge/miniforge/releases
|
# 下载 Miniforge: https://github.com/conda-forge/miniforge/releases
|
||||||
|
|
||||||
# 2. 创建 Conda 环境
|
# 2. 创建 Conda 环境
|
||||||
mamba create -n unilab python=3.11.14
|
mamba create -n unilab python=3.11.11
|
||||||
|
|
||||||
# 3. 激活环境
|
# 3. 激活环境
|
||||||
mamba activate unilab
|
mamba activate unilab
|
||||||
|
|
||||||
# 4. 安装 Uni-Lab-OS(选择其一)
|
# 4. 安装 Uni-Lab-OS
|
||||||
|
|
||||||
# 方案 A:标准安装(推荐大多数用户)
|
|
||||||
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
||||||
|
|
||||||
# 方案 B:开发者环境(可编辑模式开发)
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
pip install -e /path/to/Uni-Lab-OS # 可编辑安装
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt # 安装 pip 依赖
|
|
||||||
|
|
||||||
# 方案 C:完整版(仿真/可视化)
|
|
||||||
mamba install uni-lab::unilabos-full -c robostack-staging -c conda-forge
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**选择建议:**
|
|
||||||
- **日常使用/生产部署**:使用 `unilabos`(推荐),完整功能,开箱即用
|
|
||||||
- **开发者**:使用 `unilabos-env` + `pip install -e .` + `uv pip install -r unilabos/utils/requirements.txt`,代码修改立即生效
|
|
||||||
- **仿真/可视化**:使用 `unilabos-full`,含 Gazebo、rviz2、MoveIt
|
|
||||||
|
|
||||||
#### 1.2 验证安装
|
#### 1.2 验证安装
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -439,9 +416,6 @@ unilab --ak your_ak --sk your_sk -g test/experiments/mock_devices/mock_all.json
|
|||||||
1. 访问 Web 界面,进入"仪器耗材"模块
|
1. 访问 Web 界面,进入"仪器耗材"模块
|
||||||
2. 在"仪器设备"区域找到并添加上述设备
|
2. 在"仪器设备"区域找到并添加上述设备
|
||||||
3. 在"物料耗材"区域找到并添加容器
|
3. 在"物料耗材"区域找到并添加容器
|
||||||
4. 在workstation中配置protocol_type包含PumpTransferProtocol
|
|
||||||
|
|
||||||

|
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
@@ -794,43 +768,7 @@ Waiting for host service...
|
|||||||
|
|
||||||
详细的设备驱动编写指南请参考 [添加设备驱动](../developer_guide/add_device.md)。
|
详细的设备驱动编写指南请参考 [添加设备驱动](../developer_guide/add_device.md)。
|
||||||
|
|
||||||
#### 9.1 开发环境准备
|
#### 9.1 为什么需要自定义设备?
|
||||||
|
|
||||||
**推荐使用 `unilabos-env` + `pip install -e .` + `uv pip install`** 进行设备开发:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# 1. 创建环境并安装 unilabos-env(ROS2 + conda 依赖 + uv)
|
|
||||||
mamba create -n unilab python=3.11.14
|
|
||||||
conda activate unilab
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
|
|
||||||
# 2. 克隆代码
|
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
|
||||||
cd Uni-Lab-OS
|
|
||||||
|
|
||||||
# 3. 以可编辑模式安装(推荐使用脚本,自动检测中文环境)
|
|
||||||
python scripts/dev_install.py
|
|
||||||
|
|
||||||
# 或手动安装:
|
|
||||||
pip install -e .
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
```
|
|
||||||
|
|
||||||
**为什么使用这种方式?**
|
|
||||||
- `unilabos-env` 提供 ROS2 核心组件和 uv(通过 conda 安装,避免编译)
|
|
||||||
- `unilabos/utils/requirements.txt` 包含所有运行时需要的 pip 依赖
|
|
||||||
- `dev_install.py` 自动检测中文环境,中文系统自动使用清华镜像
|
|
||||||
- 使用 `uv` 替代 `pip`,安装速度更快
|
|
||||||
- 可编辑模式:代码修改**立即生效**,无需重新安装
|
|
||||||
|
|
||||||
**如果安装失败或速度太慢**,可以手动执行(使用清华镜像):
|
|
||||||
|
|
||||||
```bash
|
|
||||||
pip install -e . -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
|
||||||
```
|
|
||||||
|
|
||||||
#### 9.2 为什么需要自定义设备?
|
|
||||||
|
|
||||||
Uni-Lab-OS 内置了常见设备,但您的实验室可能有特殊设备需要集成:
|
Uni-Lab-OS 内置了常见设备,但您的实验室可能有特殊设备需要集成:
|
||||||
|
|
||||||
@@ -839,7 +777,7 @@ Uni-Lab-OS 内置了常见设备,但您的实验室可能有特殊设备需要
|
|||||||
- 特殊的实验流程
|
- 特殊的实验流程
|
||||||
- 第三方设备集成
|
- 第三方设备集成
|
||||||
|
|
||||||
#### 9.3 创建 Python 包
|
#### 9.2 创建 Python 包
|
||||||
|
|
||||||
为了方便开发和管理,建议为您的实验室创建独立的 Python 包。
|
为了方便开发和管理,建议为您的实验室创建独立的 Python 包。
|
||||||
|
|
||||||
@@ -876,7 +814,7 @@ touch my_lab_devices/my_lab_devices/__init__.py
|
|||||||
touch my_lab_devices/my_lab_devices/devices/__init__.py
|
touch my_lab_devices/my_lab_devices/devices/__init__.py
|
||||||
```
|
```
|
||||||
|
|
||||||
#### 9.4 创建 setup.py
|
#### 9.3 创建 setup.py
|
||||||
|
|
||||||
```python
|
```python
|
||||||
# my_lab_devices/setup.py
|
# my_lab_devices/setup.py
|
||||||
@@ -907,7 +845,7 @@ setup(
|
|||||||
)
|
)
|
||||||
```
|
```
|
||||||
|
|
||||||
#### 9.5 开发安装
|
#### 9.4 开发安装
|
||||||
|
|
||||||
使用 `-e` 参数进行可编辑安装,这样代码修改后立即生效:
|
使用 `-e` 参数进行可编辑安装,这样代码修改后立即生效:
|
||||||
|
|
||||||
@@ -922,7 +860,7 @@ pip install -e . -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
|||||||
- 方便调试和测试
|
- 方便调试和测试
|
||||||
- 支持版本控制(git)
|
- 支持版本控制(git)
|
||||||
|
|
||||||
#### 9.6 编写设备驱动
|
#### 9.5 编写设备驱动
|
||||||
|
|
||||||
创建设备驱动文件:
|
创建设备驱动文件:
|
||||||
|
|
||||||
@@ -1063,7 +1001,7 @@ class MyPump:
|
|||||||
- **返回 Dict**:所有动作方法返回字典类型
|
- **返回 Dict**:所有动作方法返回字典类型
|
||||||
- **文档字符串**:详细说明参数和功能
|
- **文档字符串**:详细说明参数和功能
|
||||||
|
|
||||||
#### 9.7 测试设备驱动
|
#### 9.6 测试设备驱动
|
||||||
|
|
||||||
创建简单的测试脚本:
|
创建简单的测试脚本:
|
||||||
|
|
||||||
|
|||||||
Binary file not shown.
|
Before Width: | Height: | Size: 81 KiB |
@@ -13,26 +13,15 @@
|
|||||||
- 开发者需要 Git 和基本的 Python 开发知识
|
- 开发者需要 Git 和基本的 Python 开发知识
|
||||||
- 自定义 msgs 需要 GitHub 账号
|
- 自定义 msgs 需要 GitHub 账号
|
||||||
|
|
||||||
## 安装包选择
|
|
||||||
|
|
||||||
Uni-Lab-OS 提供三个安装包版本,根据您的需求选择:
|
|
||||||
|
|
||||||
| 安装包 | 适用场景 | 包含组件 | 磁盘占用 |
|
|
||||||
|--------|----------|----------|----------|
|
|
||||||
| **unilabos** | **推荐大多数用户**,生产部署 | 完整安装包,开箱即用 | ~2-3 GB |
|
|
||||||
| **unilabos-env** | 开发者环境(可编辑安装) | 仅环境依赖,通过 pip 安装 unilabos | ~2 GB |
|
|
||||||
| **unilabos-full** | 仿真可视化、完整功能体验 | unilabos + 完整 ROS2 桌面版 + Gazebo + MoveIt | ~8-10 GB |
|
|
||||||
|
|
||||||
## 安装方式选择
|
## 安装方式选择
|
||||||
|
|
||||||
根据您的使用场景,选择合适的安装方式:
|
根据您的使用场景,选择合适的安装方式:
|
||||||
|
|
||||||
| 安装方式 | 适用人群 | 推荐安装包 | 特点 | 安装时间 |
|
| 安装方式 | 适用人群 | 特点 | 安装时间 |
|
||||||
| ---------------------- | -------------------- | ----------------- | ------------------------------ | ---------------------------- |
|
| ---------------------- | -------------------- | ------------------------------ | ---------------------------- |
|
||||||
| **方式一:一键安装** | 快速体验、演示 | 预打包环境 | 离线可用,无需配置 | 5-10 分钟 (网络良好的情况下) |
|
| **方式一:一键安装** | 实验室用户、快速体验 | 预打包环境,离线可用,无需配置 | 5-10 分钟 (网络良好的情况下) |
|
||||||
| **方式二:手动安装** | **大多数用户** | `unilabos` | 完整功能,开箱即用 | 10-20 分钟 |
|
| **方式二:手动安装** | 标准用户、生产环境 | 灵活配置,版本可控 | 10-20 分钟 |
|
||||||
| **方式三:开发者安装** | 开发者、需要修改源码 | `unilabos-env` | 可编辑模式,支持自定义开发 | 20-30 分钟 |
|
| **方式三:开发者安装** | 开发者、需要修改源码 | 可编辑模式,支持自定义 msgs | 20-30 分钟 |
|
||||||
| **仿真/可视化** | 仿真测试、可视化调试 | `unilabos-full` | 含 Gazebo、rviz2、MoveIt | 30-60 分钟 |
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
@@ -155,38 +144,17 @@ bash Miniforge3-$(uname)-$(uname -m).sh
|
|||||||
使用以下命令创建 Uni-Lab 专用环境:
|
使用以下命令创建 Uni-Lab 专用环境:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
mamba create -n unilab python=3.11.14 # 目前ros2组件依赖版本大多为3.11.14
|
mamba create -n unilab python=3.11.11 # 目前ros2组件依赖版本大多为3.11.11
|
||||||
mamba activate unilab
|
mamba activate unilab
|
||||||
|
mamba install -n unilab uni-lab::unilabos -c robostack-staging -c conda-forge
|
||||||
# 选择安装包(三选一):
|
|
||||||
|
|
||||||
# 方案 A:标准安装(推荐大多数用户)
|
|
||||||
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
|
||||||
|
|
||||||
# 方案 B:开发者环境(可编辑模式开发)
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
# 然后安装 unilabos 和 pip 依赖:
|
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git && cd Uni-Lab-OS
|
|
||||||
pip install -e .
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
|
|
||||||
# 方案 C:完整版(含仿真和可视化工具)
|
|
||||||
mamba install uni-lab::unilabos-full -c robostack-staging -c conda-forge
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**参数说明**:
|
**参数说明**:
|
||||||
|
|
||||||
- `-n unilab`: 创建名为 "unilab" 的环境
|
- `-n unilab`: 创建名为 "unilab" 的环境
|
||||||
- `uni-lab::unilabos`: 安装 unilabos 完整包,开箱即用(推荐)
|
- `uni-lab::unilabos`: 从 uni-lab channel 安装 unilabos 包
|
||||||
- `uni-lab::unilabos-env`: 仅安装环境依赖,适合开发者使用 `pip install -e .`
|
|
||||||
- `uni-lab::unilabos-full`: 安装完整包(含 ROS2 Desktop、Gazebo、MoveIt 等)
|
|
||||||
- `-c robostack-staging -c conda-forge`: 添加额外的软件源
|
- `-c robostack-staging -c conda-forge`: 添加额外的软件源
|
||||||
|
|
||||||
**包选择建议**:
|
|
||||||
- **日常使用/生产部署**:安装 `unilabos`(推荐,完整功能,开箱即用)
|
|
||||||
- **开发者**:安装 `unilabos-env`,然后使用 `uv pip install -r unilabos/utils/requirements.txt` 安装依赖,再 `pip install -e .` 进行可编辑安装
|
|
||||||
- **仿真/可视化**:安装 `unilabos-full`(Gazebo、rviz2、MoveIt)
|
|
||||||
|
|
||||||
**如果遇到网络问题**,可以使用清华镜像源加速下载:
|
**如果遇到网络问题**,可以使用清华镜像源加速下载:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
@@ -195,14 +163,8 @@ mamba config --add channels https://mirrors.tuna.tsinghua.edu.cn/anaconda/pkgs/m
|
|||||||
mamba config --add channels https://mirrors.tuna.tsinghua.edu.cn/anaconda/pkgs/free/
|
mamba config --add channels https://mirrors.tuna.tsinghua.edu.cn/anaconda/pkgs/free/
|
||||||
mamba config --add channels https://mirrors.tuna.tsinghua.edu.cn/anaconda/cloud/conda-forge/
|
mamba config --add channels https://mirrors.tuna.tsinghua.edu.cn/anaconda/cloud/conda-forge/
|
||||||
|
|
||||||
# 然后重新执行安装命令(推荐标准安装)
|
# 然后重新执行安装命令
|
||||||
mamba create -n unilab uni-lab::unilabos -c robostack-staging
|
mamba create -n unilab uni-lab::unilabos -c robostack-staging
|
||||||
|
|
||||||
# 或完整版(仿真/可视化)
|
|
||||||
mamba create -n unilab uni-lab::unilabos-full -c robostack-staging
|
|
||||||
|
|
||||||
# pip 安装时使用清华镜像(开发者安装时使用)
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### 第三步:激活环境
|
### 第三步:激活环境
|
||||||
@@ -241,87 +203,58 @@ cd Uni-Lab-OS
|
|||||||
cd Uni-Lab-OS
|
cd Uni-Lab-OS
|
||||||
```
|
```
|
||||||
|
|
||||||
### 第二步:安装开发环境(unilabos-env)
|
### 第二步:安装基础环境
|
||||||
|
|
||||||
**重要**:开发者请使用 `unilabos-env` 包,它专为开发者设计:
|
**推荐方式**:先通过**方式一(一键安装)**或**方式二(手动安装)**完成基础环境的安装,这将包含所有必需的依赖项(ROS2、msgs 等)。
|
||||||
- 包含 ROS2 核心组件和消息包(ros-humble-ros-core、std-msgs、geometry-msgs 等)
|
|
||||||
- 包含 transforms3d、cv-bridge、tf2 等 conda 依赖
|
#### 选项 A:通过一键安装(推荐)
|
||||||
- 包含 `uv` 工具,用于快速安装 pip 依赖
|
|
||||||
- **不包含** pip 依赖和 unilabos 包(由 `pip install -e .` 和 `uv pip install` 安装)
|
参考上文"方式一:一键安装",完成基础环境的安装后,激活环境:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# 创建并激活环境
|
|
||||||
mamba create -n unilab python=3.11.14
|
|
||||||
conda activate unilab
|
conda activate unilab
|
||||||
|
|
||||||
# 安装开发者环境包(ROS2 + conda 依赖 + uv)
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
```
|
```
|
||||||
|
|
||||||
### 第三步:安装 pip 依赖和可编辑模式安装
|
#### 选项 B:通过手动安装
|
||||||
|
|
||||||
克隆代码并安装依赖:
|
参考上文"方式二:手动安装",创建并安装环境:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
mamba create -n unilab python=3.11.11
|
||||||
|
conda activate unilab
|
||||||
|
mamba install -n unilab uni-lab::unilabos -c robostack-staging -c conda-forge
|
||||||
|
```
|
||||||
|
|
||||||
|
**说明**:这会安装包括 Python 3.11.11、ROS2 Humble、ros-humble-unilabos-msgs 和所有必需依赖
|
||||||
|
|
||||||
|
### 第三步:切换到开发版本
|
||||||
|
|
||||||
|
现在你已经有了一个完整可用的 Uni-Lab 环境,接下来将 unilabos 包切换为开发版本:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# 确保环境已激活
|
# 确保环境已激活
|
||||||
conda activate unilab
|
conda activate unilab
|
||||||
|
|
||||||
# 克隆仓库(如果还未克隆)
|
# 卸载 pip 安装的 unilabos(保留所有 conda 依赖)
|
||||||
git clone https://github.com/deepmodeling/Uni-Lab-OS.git
|
pip uninstall unilabos -y
|
||||||
cd Uni-Lab-OS
|
|
||||||
|
|
||||||
# 切换到 dev 分支(可选)
|
# 克隆 dev 分支(如果还未克隆)
|
||||||
|
cd /path/to/your/workspace
|
||||||
|
git clone -b dev https://github.com/deepmodeling/Uni-Lab-OS.git
|
||||||
|
# 或者如果已经克隆,切换到 dev 分支
|
||||||
|
cd Uni-Lab-OS
|
||||||
git checkout dev
|
git checkout dev
|
||||||
git pull
|
git pull
|
||||||
```
|
|
||||||
|
|
||||||
**推荐:使用安装脚本**(自动检测中文环境,使用 uv 加速):
|
# 以可编辑模式安装开发版 unilabos
|
||||||
|
|
||||||
```bash
|
|
||||||
# 自动检测中文环境,如果是中文系统则使用清华镜像
|
|
||||||
python scripts/dev_install.py
|
|
||||||
|
|
||||||
# 或者手动指定:
|
|
||||||
python scripts/dev_install.py --china # 强制使用清华镜像
|
|
||||||
python scripts/dev_install.py --no-mirror # 强制使用 PyPI
|
|
||||||
python scripts/dev_install.py --skip-deps # 跳过 pip 依赖安装
|
|
||||||
python scripts/dev_install.py --use-pip # 使用 pip 而非 uv
|
|
||||||
```
|
|
||||||
|
|
||||||
**手动安装**(如果脚本安装失败或速度太慢):
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# 1. 安装 unilabos(可编辑模式)
|
|
||||||
pip install -e .
|
|
||||||
|
|
||||||
# 2. 使用 uv 安装 pip 依赖(推荐,速度更快)
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
|
|
||||||
# 国内用户使用清华镜像:
|
|
||||||
pip install -e . -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
pip install -e . -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
||||||
uv pip install -r unilabos/utils/requirements.txt -i https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**注意**:
|
**参数说明**:
|
||||||
- `uv` 已包含在 `unilabos-env` 中,无需单独安装
|
|
||||||
- `unilabos/utils/requirements.txt` 包含运行 unilabos 所需的所有 pip 依赖
|
|
||||||
- 部分特殊包(如 pylabrobot)会在运行时由 unilabos 自动检测并安装
|
|
||||||
|
|
||||||
**为什么使用可编辑模式?**
|
- `-e`: editable mode(可编辑模式),代码修改立即生效,无需重新安装
|
||||||
|
- `-i`: 使用清华镜像源加速下载
|
||||||
- `-e` (editable mode):代码修改**立即生效**,无需重新安装
|
- `pip uninstall unilabos`: 只卸载 pip 安装的 unilabos 包,不影响 conda 安装的其他依赖(如 ROS2、msgs 等)
|
||||||
- 适合开发调试:修改代码后直接运行测试
|
|
||||||
- 与 `unilabos-env` 配合:环境依赖由 conda 管理,unilabos 代码由 pip 管理
|
|
||||||
|
|
||||||
**验证安装**:
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# 检查 unilabos 版本
|
|
||||||
python -c "import unilabos; print(unilabos.__version__)"
|
|
||||||
|
|
||||||
# 检查安装位置(应该指向你的代码目录)
|
|
||||||
pip show unilabos | grep Location
|
|
||||||
```
|
|
||||||
|
|
||||||
### 第四步:安装或自定义 ros-humble-unilabos-msgs(可选)
|
### 第四步:安装或自定义 ros-humble-unilabos-msgs(可选)
|
||||||
|
|
||||||
@@ -531,45 +464,7 @@ cd $CONDA_PREFIX/envs/unilab
|
|||||||
|
|
||||||
### 问题 8: 环境很大,有办法减小吗?
|
### 问题 8: 环境很大,有办法减小吗?
|
||||||
|
|
||||||
**解决方案**:
|
**解决方案**: 预打包的环境包含所有依赖,通常较大(压缩后 2-5GB)。这是为了确保离线安装和完整功能。如果空间有限,考虑使用方式二手动安装,只安装需要的组件。
|
||||||
|
|
||||||
1. **使用 `unilabos` 标准版**(推荐大多数用户):
|
|
||||||
```bash
|
|
||||||
mamba install uni-lab::unilabos -c robostack-staging -c conda-forge
|
|
||||||
```
|
|
||||||
标准版包含完整功能,环境大小约 2-3GB(相比完整版的 8-10GB)。
|
|
||||||
|
|
||||||
2. **使用 `unilabos-env` 开发者版**(最小化):
|
|
||||||
```bash
|
|
||||||
mamba install uni-lab::unilabos-env -c robostack-staging -c conda-forge
|
|
||||||
# 然后手动安装依赖
|
|
||||||
pip install -e .
|
|
||||||
uv pip install -r unilabos/utils/requirements.txt
|
|
||||||
```
|
|
||||||
开发者版只包含环境依赖,体积最小约 2GB。
|
|
||||||
|
|
||||||
3. **按需安装额外组件**:
|
|
||||||
如果后续需要特定功能,可以单独安装:
|
|
||||||
```bash
|
|
||||||
# 需要 Jupyter
|
|
||||||
mamba install jupyter jupyros
|
|
||||||
|
|
||||||
# 需要可视化
|
|
||||||
mamba install matplotlib opencv
|
|
||||||
|
|
||||||
# 需要仿真(注意:这会安装大量依赖)
|
|
||||||
mamba install ros-humble-gazebo-ros
|
|
||||||
```
|
|
||||||
|
|
||||||
4. **预打包环境问题**:
|
|
||||||
预打包环境(方式一)包含所有依赖,通常较大(压缩后 2-5GB)。这是为了确保离线安装和完整功能。
|
|
||||||
|
|
||||||
**包选择建议**:
|
|
||||||
| 需求 | 推荐包 | 预估大小 |
|
|
||||||
|------|--------|----------|
|
|
||||||
| 日常使用/生产部署 | `unilabos` | ~2-3 GB |
|
|
||||||
| 开发调试(可编辑模式) | `unilabos-env` | ~2 GB |
|
|
||||||
| 仿真/可视化 | `unilabos-full` | ~8-10 GB |
|
|
||||||
|
|
||||||
### 问题 9: 如何更新到最新版本?
|
### 问题 9: 如何更新到最新版本?
|
||||||
|
|
||||||
@@ -616,7 +511,6 @@ mamba update ros-humble-unilabos-msgs -c uni-lab -c robostack-staging -c conda-f
|
|||||||
|
|
||||||
**提示**:
|
**提示**:
|
||||||
|
|
||||||
- **大多数用户**推荐使用方式二(手动安装)的 `unilabos` 标准版
|
- 生产环境推荐使用方式二(手动安装)的稳定版本
|
||||||
- **开发者**推荐使用方式三(开发者安装),安装 `unilabos-env` 后使用 `uv pip install -r unilabos/utils/requirements.txt` 安装依赖
|
- 开发和测试推荐使用方式三(开发者安装)
|
||||||
- **仿真/可视化**推荐安装 `unilabos-full` 完整版
|
- 快速体验和演示推荐使用方式一(一键安装)
|
||||||
- **快速体验和演示**推荐使用方式一(一键安装)
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package:
|
package:
|
||||||
name: ros-humble-unilabos-msgs
|
name: ros-humble-unilabos-msgs
|
||||||
version: 0.10.17
|
version: 0.10.15
|
||||||
source:
|
source:
|
||||||
path: ../../unilabos_msgs
|
path: ../../unilabos_msgs
|
||||||
target_directory: src
|
target_directory: src
|
||||||
@@ -25,7 +25,7 @@ requirements:
|
|||||||
build:
|
build:
|
||||||
- ${{ compiler('cxx') }}
|
- ${{ compiler('cxx') }}
|
||||||
- ${{ compiler('c') }}
|
- ${{ compiler('c') }}
|
||||||
- python ==3.11.14
|
- python ==3.11.11
|
||||||
- numpy
|
- numpy
|
||||||
- if: build_platform != target_platform
|
- if: build_platform != target_platform
|
||||||
then:
|
then:
|
||||||
@@ -63,14 +63,14 @@ requirements:
|
|||||||
- robostack-staging::ros-humble-rosidl-default-generators
|
- robostack-staging::ros-humble-rosidl-default-generators
|
||||||
- robostack-staging::ros-humble-std-msgs
|
- robostack-staging::ros-humble-std-msgs
|
||||||
- robostack-staging::ros-humble-geometry-msgs
|
- robostack-staging::ros-humble-geometry-msgs
|
||||||
- robostack-staging::ros2-distro-mutex=0.7
|
- robostack-staging::ros2-distro-mutex=0.6
|
||||||
run:
|
run:
|
||||||
- robostack-staging::ros-humble-action-msgs
|
- robostack-staging::ros-humble-action-msgs
|
||||||
- robostack-staging::ros-humble-ros-workspace
|
- robostack-staging::ros-humble-ros-workspace
|
||||||
- robostack-staging::ros-humble-rosidl-default-runtime
|
- robostack-staging::ros-humble-rosidl-default-runtime
|
||||||
- robostack-staging::ros-humble-std-msgs
|
- robostack-staging::ros-humble-std-msgs
|
||||||
- robostack-staging::ros-humble-geometry-msgs
|
- robostack-staging::ros-humble-geometry-msgs
|
||||||
- robostack-staging::ros2-distro-mutex=0.7
|
- robostack-staging::ros2-distro-mutex=0.6
|
||||||
- if: osx and x86_64
|
- if: osx and x86_64
|
||||||
then:
|
then:
|
||||||
- __osx >=${{ MACOSX_DEPLOYMENT_TARGET|default('10.14') }}
|
- __osx >=${{ MACOSX_DEPLOYMENT_TARGET|default('10.14') }}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package:
|
package:
|
||||||
name: unilabos
|
name: unilabos
|
||||||
version: "0.10.17"
|
version: "0.10.15"
|
||||||
|
|
||||||
source:
|
source:
|
||||||
path: ../..
|
path: ../..
|
||||||
|
|||||||
@@ -85,7 +85,7 @@ Verification:
|
|||||||
-------------
|
-------------
|
||||||
|
|
||||||
The verify_installation.py script will check:
|
The verify_installation.py script will check:
|
||||||
- Python version (3.11.14)
|
- Python version (3.11.11)
|
||||||
- ROS2 rclpy installation
|
- ROS2 rclpy installation
|
||||||
- UniLabOS installation and dependencies
|
- UniLabOS installation and dependencies
|
||||||
|
|
||||||
@@ -104,7 +104,7 @@ Build Information:
|
|||||||
|
|
||||||
Branch: {branch}
|
Branch: {branch}
|
||||||
Platform: {platform}
|
Platform: {platform}
|
||||||
Python: 3.11.14
|
Python: 3.11.11
|
||||||
Date: {build_date}
|
Date: {build_date}
|
||||||
|
|
||||||
Troubleshooting:
|
Troubleshooting:
|
||||||
|
|||||||
@@ -1,214 +0,0 @@
|
|||||||
#!/usr/bin/env python3
|
|
||||||
"""
|
|
||||||
Development installation script for UniLabOS.
|
|
||||||
Auto-detects Chinese locale and uses appropriate mirror.
|
|
||||||
|
|
||||||
Usage:
|
|
||||||
python scripts/dev_install.py
|
|
||||||
python scripts/dev_install.py --no-mirror # Force no mirror
|
|
||||||
python scripts/dev_install.py --china # Force China mirror
|
|
||||||
python scripts/dev_install.py --skip-deps # Skip pip dependencies installation
|
|
||||||
|
|
||||||
Flow:
|
|
||||||
1. pip install -e . (install unilabos in editable mode)
|
|
||||||
2. Detect Chinese locale
|
|
||||||
3. Use uv to install pip dependencies from requirements.txt
|
|
||||||
4. Special packages (like pylabrobot) are handled by environment_check.py at runtime
|
|
||||||
"""
|
|
||||||
|
|
||||||
import locale
|
|
||||||
import subprocess
|
|
||||||
import sys
|
|
||||||
import argparse
|
|
||||||
from pathlib import Path
|
|
||||||
|
|
||||||
# Tsinghua mirror URL
|
|
||||||
TSINGHUA_MIRROR = "https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple"
|
|
||||||
|
|
||||||
|
|
||||||
def is_chinese_locale() -> bool:
|
|
||||||
"""
|
|
||||||
Detect if system is in Chinese locale.
|
|
||||||
Same logic as EnvironmentChecker._is_chinese_locale()
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
lang = locale.getdefaultlocale()[0]
|
|
||||||
if lang and ("zh" in lang.lower() or "chinese" in lang.lower()):
|
|
||||||
return True
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def run_command(cmd: list, description: str, retry: int = 2) -> bool:
|
|
||||||
"""Run command with retry support."""
|
|
||||||
print(f"[INFO] {description}")
|
|
||||||
print(f"[CMD] {' '.join(cmd)}")
|
|
||||||
|
|
||||||
for attempt in range(retry + 1):
|
|
||||||
try:
|
|
||||||
result = subprocess.run(cmd, check=True, timeout=600)
|
|
||||||
print(f"[OK] {description}")
|
|
||||||
return True
|
|
||||||
except subprocess.CalledProcessError as e:
|
|
||||||
if attempt < retry:
|
|
||||||
print(f"[WARN] Attempt {attempt + 1} failed, retrying...")
|
|
||||||
else:
|
|
||||||
print(f"[ERROR] {description} failed: {e}")
|
|
||||||
return False
|
|
||||||
except subprocess.TimeoutExpired:
|
|
||||||
print(f"[ERROR] {description} timed out")
|
|
||||||
return False
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def install_editable(project_root: Path, use_mirror: bool) -> bool:
|
|
||||||
"""Install unilabos in editable mode using pip."""
|
|
||||||
cmd = [sys.executable, "-m", "pip", "install", "-e", str(project_root)]
|
|
||||||
if use_mirror:
|
|
||||||
cmd.extend(["-i", TSINGHUA_MIRROR])
|
|
||||||
|
|
||||||
return run_command(cmd, "Installing unilabos in editable mode")
|
|
||||||
|
|
||||||
|
|
||||||
def install_requirements_uv(requirements_file: Path, use_mirror: bool) -> bool:
|
|
||||||
"""Install pip dependencies using uv (installed via conda-forge::uv)."""
|
|
||||||
cmd = ["uv", "pip", "install", "-r", str(requirements_file)]
|
|
||||||
if use_mirror:
|
|
||||||
cmd.extend(["-i", TSINGHUA_MIRROR])
|
|
||||||
|
|
||||||
return run_command(cmd, "Installing pip dependencies with uv", retry=2)
|
|
||||||
|
|
||||||
|
|
||||||
def install_requirements_pip(requirements_file: Path, use_mirror: bool) -> bool:
|
|
||||||
"""Fallback: Install pip dependencies using pip."""
|
|
||||||
cmd = [sys.executable, "-m", "pip", "install", "-r", str(requirements_file)]
|
|
||||||
if use_mirror:
|
|
||||||
cmd.extend(["-i", TSINGHUA_MIRROR])
|
|
||||||
|
|
||||||
return run_command(cmd, "Installing pip dependencies with pip", retry=2)
|
|
||||||
|
|
||||||
|
|
||||||
def check_uv_available() -> bool:
|
|
||||||
"""Check if uv is available (installed via conda-forge::uv)."""
|
|
||||||
try:
|
|
||||||
subprocess.run(["uv", "--version"], capture_output=True, check=True)
|
|
||||||
return True
|
|
||||||
except (subprocess.CalledProcessError, FileNotFoundError):
|
|
||||||
return False
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
|
||||||
parser = argparse.ArgumentParser(description="Development installation script for UniLabOS")
|
|
||||||
parser.add_argument("--china", action="store_true", help="Force use China mirror (Tsinghua)")
|
|
||||||
parser.add_argument("--no-mirror", action="store_true", help="Force use default PyPI (no mirror)")
|
|
||||||
parser.add_argument(
|
|
||||||
"--skip-deps", action="store_true", help="Skip pip dependencies installation (only install unilabos)"
|
|
||||||
)
|
|
||||||
parser.add_argument("--use-pip", action="store_true", help="Use pip instead of uv for dependencies")
|
|
||||||
args = parser.parse_args()
|
|
||||||
|
|
||||||
# Determine project root
|
|
||||||
script_dir = Path(__file__).parent
|
|
||||||
project_root = script_dir.parent
|
|
||||||
requirements_file = project_root / "unilabos" / "utils" / "requirements.txt"
|
|
||||||
|
|
||||||
if not (project_root / "setup.py").exists():
|
|
||||||
print(f"[ERROR] setup.py not found in {project_root}")
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
print("=" * 60)
|
|
||||||
print("UniLabOS Development Installation")
|
|
||||||
print("=" * 60)
|
|
||||||
print(f"Project root: {project_root}")
|
|
||||||
print()
|
|
||||||
|
|
||||||
# Determine mirror usage based on locale
|
|
||||||
if args.no_mirror:
|
|
||||||
use_mirror = False
|
|
||||||
print("[INFO] Mirror disabled by --no-mirror flag")
|
|
||||||
elif args.china:
|
|
||||||
use_mirror = True
|
|
||||||
print("[INFO] China mirror enabled by --china flag")
|
|
||||||
else:
|
|
||||||
use_mirror = is_chinese_locale()
|
|
||||||
if use_mirror:
|
|
||||||
print("[INFO] Chinese locale detected, using Tsinghua mirror")
|
|
||||||
else:
|
|
||||||
print("[INFO] Non-Chinese locale detected, using default PyPI")
|
|
||||||
|
|
||||||
print()
|
|
||||||
|
|
||||||
# Step 1: Install unilabos in editable mode
|
|
||||||
print("[STEP 1] Installing unilabos in editable mode...")
|
|
||||||
if not install_editable(project_root, use_mirror):
|
|
||||||
print("[ERROR] Failed to install unilabos")
|
|
||||||
print()
|
|
||||||
print("Manual fallback:")
|
|
||||||
if use_mirror:
|
|
||||||
print(f" pip install -e {project_root} -i {TSINGHUA_MIRROR}")
|
|
||||||
else:
|
|
||||||
print(f" pip install -e {project_root}")
|
|
||||||
sys.exit(1)
|
|
||||||
|
|
||||||
print()
|
|
||||||
|
|
||||||
# Step 2: Install pip dependencies
|
|
||||||
if args.skip_deps:
|
|
||||||
print("[INFO] Skipping pip dependencies installation (--skip-deps)")
|
|
||||||
else:
|
|
||||||
print("[STEP 2] Installing pip dependencies...")
|
|
||||||
|
|
||||||
if not requirements_file.exists():
|
|
||||||
print(f"[WARN] Requirements file not found: {requirements_file}")
|
|
||||||
print("[INFO] Skipping dependencies installation")
|
|
||||||
else:
|
|
||||||
# Try uv first (faster), fallback to pip
|
|
||||||
if args.use_pip:
|
|
||||||
print("[INFO] Using pip (--use-pip flag)")
|
|
||||||
success = install_requirements_pip(requirements_file, use_mirror)
|
|
||||||
elif check_uv_available():
|
|
||||||
print("[INFO] Using uv (installed via conda-forge::uv)")
|
|
||||||
success = install_requirements_uv(requirements_file, use_mirror)
|
|
||||||
if not success:
|
|
||||||
print("[WARN] uv failed, falling back to pip...")
|
|
||||||
success = install_requirements_pip(requirements_file, use_mirror)
|
|
||||||
else:
|
|
||||||
print("[WARN] uv not available (should be installed via: mamba install conda-forge::uv)")
|
|
||||||
print("[INFO] Falling back to pip...")
|
|
||||||
success = install_requirements_pip(requirements_file, use_mirror)
|
|
||||||
|
|
||||||
if not success:
|
|
||||||
print()
|
|
||||||
print("[WARN] Failed to install some dependencies automatically.")
|
|
||||||
print("You can manually install them:")
|
|
||||||
if use_mirror:
|
|
||||||
print(f" uv pip install -r {requirements_file} -i {TSINGHUA_MIRROR}")
|
|
||||||
print(" or:")
|
|
||||||
print(f" pip install -r {requirements_file} -i {TSINGHUA_MIRROR}")
|
|
||||||
else:
|
|
||||||
print(f" uv pip install -r {requirements_file}")
|
|
||||||
print(" or:")
|
|
||||||
print(f" pip install -r {requirements_file}")
|
|
||||||
|
|
||||||
print()
|
|
||||||
print("=" * 60)
|
|
||||||
print("Installation complete!")
|
|
||||||
print("=" * 60)
|
|
||||||
print()
|
|
||||||
print("Note: Some special packages (like pylabrobot) are installed")
|
|
||||||
print("automatically at runtime by unilabos if needed.")
|
|
||||||
print()
|
|
||||||
print("Verify installation:")
|
|
||||||
print(' python -c "import unilabos; print(unilabos.__version__)"')
|
|
||||||
print()
|
|
||||||
print("If you encounter issues, you can manually install dependencies:")
|
|
||||||
if use_mirror:
|
|
||||||
print(f" uv pip install -r unilabos/utils/requirements.txt -i {TSINGHUA_MIRROR}")
|
|
||||||
else:
|
|
||||||
print(" uv pip install -r unilabos/utils/requirements.txt")
|
|
||||||
print()
|
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
||||||
@@ -2,7 +2,6 @@ import json
|
|||||||
import logging
|
import logging
|
||||||
import traceback
|
import traceback
|
||||||
import uuid
|
import uuid
|
||||||
import xml.etree.ElementTree as ET
|
|
||||||
from typing import Any, Dict, List
|
from typing import Any, Dict, List
|
||||||
|
|
||||||
import networkx as nx
|
import networkx as nx
|
||||||
@@ -25,7 +24,15 @@ class SimpleGraph:
|
|||||||
|
|
||||||
def add_edge(self, source, target, **attrs):
|
def add_edge(self, source, target, **attrs):
|
||||||
"""添加边"""
|
"""添加边"""
|
||||||
edge = {"source": source, "target": target, **attrs}
|
# edge = {"source": source, "target": target, **attrs}
|
||||||
|
edge = {
|
||||||
|
"source": source, "target": target,
|
||||||
|
"source_node_uuid": source,
|
||||||
|
"target_node_uuid": target,
|
||||||
|
"source_handle_io": "source",
|
||||||
|
"target_handle_io": "target",
|
||||||
|
**attrs
|
||||||
|
}
|
||||||
self.edges.append(edge)
|
self.edges.append(edge)
|
||||||
|
|
||||||
def to_dict(self):
|
def to_dict(self):
|
||||||
@@ -42,6 +49,7 @@ class SimpleGraph:
|
|||||||
"multigraph": False,
|
"multigraph": False,
|
||||||
"graph": {},
|
"graph": {},
|
||||||
"nodes": nodes_list,
|
"nodes": nodes_list,
|
||||||
|
"edges": self.edges,
|
||||||
"links": self.edges,
|
"links": self.edges,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -58,495 +66,8 @@ def extract_json_from_markdown(text: str) -> str:
|
|||||||
return text
|
return text
|
||||||
|
|
||||||
|
|
||||||
def convert_to_type(val: str) -> Any:
|
|
||||||
"""将字符串值转换为适当的数据类型"""
|
|
||||||
if val == "True":
|
|
||||||
return True
|
|
||||||
if val == "False":
|
|
||||||
return False
|
|
||||||
if val == "?":
|
|
||||||
return None
|
|
||||||
if val.endswith(" g"):
|
|
||||||
return float(val.split(" ")[0])
|
|
||||||
if val.endswith("mg"):
|
|
||||||
return float(val.split("mg")[0])
|
|
||||||
elif val.endswith("mmol"):
|
|
||||||
return float(val.split("mmol")[0]) / 1000
|
|
||||||
elif val.endswith("mol"):
|
|
||||||
return float(val.split("mol")[0])
|
|
||||||
elif val.endswith("ml"):
|
|
||||||
return float(val.split("ml")[0])
|
|
||||||
elif val.endswith("RPM"):
|
|
||||||
return float(val.split("RPM")[0])
|
|
||||||
elif val.endswith(" °C"):
|
|
||||||
return float(val.split(" ")[0])
|
|
||||||
elif val.endswith(" %"):
|
|
||||||
return float(val.split(" ")[0])
|
|
||||||
return val
|
|
||||||
|
|
||||||
|
|
||||||
def refactor_data(data: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
|
||||||
"""统一的数据重构函数,根据操作类型自动选择模板"""
|
|
||||||
refactored_data = []
|
|
||||||
|
|
||||||
# 定义操作映射,包含生物实验和有机化学的所有操作
|
|
||||||
OPERATION_MAPPING = {
|
|
||||||
# 生物实验操作
|
|
||||||
"transfer_liquid": "SynBioFactory-liquid_handler.prcxi-transfer_liquid",
|
|
||||||
"transfer": "SynBioFactory-liquid_handler.biomek-transfer",
|
|
||||||
"incubation": "SynBioFactory-liquid_handler.biomek-incubation",
|
|
||||||
"move_labware": "SynBioFactory-liquid_handler.biomek-move_labware",
|
|
||||||
"oscillation": "SynBioFactory-liquid_handler.biomek-oscillation",
|
|
||||||
# 有机化学操作
|
|
||||||
"HeatChillToTemp": "SynBioFactory-workstation-HeatChillProtocol",
|
|
||||||
"StopHeatChill": "SynBioFactory-workstation-HeatChillStopProtocol",
|
|
||||||
"StartHeatChill": "SynBioFactory-workstation-HeatChillStartProtocol",
|
|
||||||
"HeatChill": "SynBioFactory-workstation-HeatChillProtocol",
|
|
||||||
"Dissolve": "SynBioFactory-workstation-DissolveProtocol",
|
|
||||||
"Transfer": "SynBioFactory-workstation-TransferProtocol",
|
|
||||||
"Evaporate": "SynBioFactory-workstation-EvaporateProtocol",
|
|
||||||
"Recrystallize": "SynBioFactory-workstation-RecrystallizeProtocol",
|
|
||||||
"Filter": "SynBioFactory-workstation-FilterProtocol",
|
|
||||||
"Dry": "SynBioFactory-workstation-DryProtocol",
|
|
||||||
"Add": "SynBioFactory-workstation-AddProtocol",
|
|
||||||
}
|
|
||||||
|
|
||||||
UNSUPPORTED_OPERATIONS = ["Purge", "Wait", "Stir", "ResetHandling"]
|
|
||||||
|
|
||||||
for step in data:
|
|
||||||
operation = step.get("action")
|
|
||||||
if not operation or operation in UNSUPPORTED_OPERATIONS:
|
|
||||||
continue
|
|
||||||
|
|
||||||
# 处理重复操作
|
|
||||||
if operation == "Repeat":
|
|
||||||
times = step.get("times", step.get("parameters", {}).get("times", 1))
|
|
||||||
sub_steps = step.get("steps", step.get("parameters", {}).get("steps", []))
|
|
||||||
for i in range(int(times)):
|
|
||||||
sub_data = refactor_data(sub_steps)
|
|
||||||
refactored_data.extend(sub_data)
|
|
||||||
continue
|
|
||||||
|
|
||||||
# 获取模板名称
|
|
||||||
template = OPERATION_MAPPING.get(operation)
|
|
||||||
if not template:
|
|
||||||
# 自动推断模板类型
|
|
||||||
if operation.lower() in ["transfer", "incubation", "move_labware", "oscillation"]:
|
|
||||||
template = f"SynBioFactory-liquid_handler.biomek-{operation}"
|
|
||||||
else:
|
|
||||||
template = f"SynBioFactory-workstation-{operation}Protocol"
|
|
||||||
|
|
||||||
# 创建步骤数据
|
|
||||||
step_data = {
|
|
||||||
"template": template,
|
|
||||||
"description": step.get("description", step.get("purpose", f"{operation} operation")),
|
|
||||||
"lab_node_type": "Device",
|
|
||||||
"parameters": step.get("parameters", step.get("action_args", {})),
|
|
||||||
}
|
|
||||||
refactored_data.append(step_data)
|
|
||||||
|
|
||||||
return refactored_data
|
|
||||||
|
|
||||||
|
|
||||||
def build_protocol_graph(
|
|
||||||
labware_info: List[Dict[str, Any]], protocol_steps: List[Dict[str, Any]], workstation_name: str
|
|
||||||
) -> SimpleGraph:
|
|
||||||
"""统一的协议图构建函数,根据设备类型自动选择构建逻辑"""
|
|
||||||
G = SimpleGraph()
|
|
||||||
resource_last_writer = {}
|
|
||||||
LAB_NAME = "SynBioFactory"
|
|
||||||
|
|
||||||
protocol_steps = refactor_data(protocol_steps)
|
|
||||||
|
|
||||||
# 检查协议步骤中的模板来判断协议类型
|
|
||||||
has_biomek_template = any(
|
|
||||||
("biomek" in step.get("template", "")) or ("prcxi" in step.get("template", ""))
|
|
||||||
for step in protocol_steps
|
|
||||||
)
|
|
||||||
|
|
||||||
if has_biomek_template:
|
|
||||||
# 生物实验协议图构建
|
|
||||||
for labware_id, labware in labware_info.items():
|
|
||||||
node_id = str(uuid.uuid4())
|
|
||||||
|
|
||||||
labware_attrs = labware.copy()
|
|
||||||
labware_id = labware_attrs.pop("id", labware_attrs.get("name", f"labware_{uuid.uuid4()}"))
|
|
||||||
labware_attrs["description"] = labware_id
|
|
||||||
labware_attrs["lab_node_type"] = (
|
|
||||||
"Reagent" if "Plate" in str(labware_id) else "Labware" if "Rack" in str(labware_id) else "Sample"
|
|
||||||
)
|
|
||||||
labware_attrs["device_id"] = workstation_name
|
|
||||||
|
|
||||||
G.add_node(node_id, template=f"{LAB_NAME}-host_node-create_resource", **labware_attrs)
|
|
||||||
resource_last_writer[labware_id] = f"{node_id}:labware"
|
|
||||||
|
|
||||||
# 处理协议步骤
|
|
||||||
prev_node = None
|
|
||||||
for i, step in enumerate(protocol_steps):
|
|
||||||
node_id = str(uuid.uuid4())
|
|
||||||
G.add_node(node_id, **step)
|
|
||||||
|
|
||||||
# 添加控制流边
|
|
||||||
if prev_node is not None:
|
|
||||||
G.add_edge(prev_node, node_id, source_port="ready", target_port="ready")
|
|
||||||
prev_node = node_id
|
|
||||||
|
|
||||||
# 处理物料流
|
|
||||||
params = step.get("parameters", {})
|
|
||||||
if "sources" in params and params["sources"] in resource_last_writer:
|
|
||||||
source_node, source_port = resource_last_writer[params["sources"]].split(":")
|
|
||||||
G.add_edge(source_node, node_id, source_port=source_port, target_port="labware")
|
|
||||||
|
|
||||||
if "targets" in params:
|
|
||||||
resource_last_writer[params["targets"]] = f"{node_id}:labware"
|
|
||||||
|
|
||||||
# 添加协议结束节点
|
|
||||||
end_id = str(uuid.uuid4())
|
|
||||||
G.add_node(end_id, template=f"{LAB_NAME}-liquid_handler.biomek-run_protocol")
|
|
||||||
if prev_node is not None:
|
|
||||||
G.add_edge(prev_node, end_id, source_port="ready", target_port="ready")
|
|
||||||
|
|
||||||
else:
|
|
||||||
# 有机化学协议图构建
|
|
||||||
WORKSTATION_ID = workstation_name
|
|
||||||
|
|
||||||
# 为所有labware创建资源节点
|
|
||||||
for item_id, item in labware_info.items():
|
|
||||||
# item_id = item.get("id") or item.get("name", f"item_{uuid.uuid4()}")
|
|
||||||
node_id = str(uuid.uuid4())
|
|
||||||
|
|
||||||
# 判断节点类型
|
|
||||||
if item.get("type") == "hardware" or "reactor" in str(item_id).lower():
|
|
||||||
if "reactor" not in str(item_id).lower():
|
|
||||||
continue
|
|
||||||
lab_node_type = "Sample"
|
|
||||||
description = f"Prepare Reactor: {item_id}"
|
|
||||||
liquid_type = []
|
|
||||||
liquid_volume = []
|
|
||||||
else:
|
|
||||||
lab_node_type = "Reagent"
|
|
||||||
description = f"Add Reagent to Flask: {item_id}"
|
|
||||||
liquid_type = [item_id]
|
|
||||||
liquid_volume = [1e5]
|
|
||||||
|
|
||||||
G.add_node(
|
|
||||||
node_id,
|
|
||||||
template=f"{LAB_NAME}-host_node-create_resource",
|
|
||||||
description=description,
|
|
||||||
lab_node_type=lab_node_type,
|
|
||||||
res_id=item_id,
|
|
||||||
device_id=WORKSTATION_ID,
|
|
||||||
class_name="container",
|
|
||||||
parent=WORKSTATION_ID,
|
|
||||||
bind_locations={"x": 0.0, "y": 0.0, "z": 0.0},
|
|
||||||
liquid_input_slot=[-1],
|
|
||||||
liquid_type=liquid_type,
|
|
||||||
liquid_volume=liquid_volume,
|
|
||||||
slot_on_deck="",
|
|
||||||
role=item.get("role", ""),
|
|
||||||
)
|
|
||||||
resource_last_writer[item_id] = f"{node_id}:labware"
|
|
||||||
|
|
||||||
last_control_node_id = None
|
|
||||||
|
|
||||||
# 处理协议步骤
|
|
||||||
for step in protocol_steps:
|
|
||||||
node_id = str(uuid.uuid4())
|
|
||||||
G.add_node(node_id, **step)
|
|
||||||
|
|
||||||
# 控制流
|
|
||||||
if last_control_node_id is not None:
|
|
||||||
G.add_edge(last_control_node_id, node_id, source_port="ready", target_port="ready")
|
|
||||||
last_control_node_id = node_id
|
|
||||||
|
|
||||||
# 物料流
|
|
||||||
params = step.get("parameters", {})
|
|
||||||
input_resources = {
|
|
||||||
"Vessel": params.get("vessel"),
|
|
||||||
"ToVessel": params.get("to_vessel"),
|
|
||||||
"FromVessel": params.get("from_vessel"),
|
|
||||||
"reagent": params.get("reagent"),
|
|
||||||
"solvent": params.get("solvent"),
|
|
||||||
"compound": params.get("compound"),
|
|
||||||
"sources": params.get("sources"),
|
|
||||||
"targets": params.get("targets"),
|
|
||||||
}
|
|
||||||
|
|
||||||
for target_port, resource_name in input_resources.items():
|
|
||||||
if resource_name and resource_name in resource_last_writer:
|
|
||||||
source_node, source_port = resource_last_writer[resource_name].split(":")
|
|
||||||
G.add_edge(source_node, node_id, source_port=source_port, target_port=target_port)
|
|
||||||
|
|
||||||
output_resources = {
|
|
||||||
"VesselOut": params.get("vessel"),
|
|
||||||
"FromVesselOut": params.get("from_vessel"),
|
|
||||||
"ToVesselOut": params.get("to_vessel"),
|
|
||||||
"FiltrateOut": params.get("filtrate_vessel"),
|
|
||||||
"reagent": params.get("reagent"),
|
|
||||||
"solvent": params.get("solvent"),
|
|
||||||
"compound": params.get("compound"),
|
|
||||||
"sources_out": params.get("sources"),
|
|
||||||
"targets_out": params.get("targets"),
|
|
||||||
}
|
|
||||||
|
|
||||||
for source_port, resource_name in output_resources.items():
|
|
||||||
if resource_name:
|
|
||||||
resource_last_writer[resource_name] = f"{node_id}:{source_port}"
|
|
||||||
|
|
||||||
return G
|
|
||||||
|
|
||||||
|
|
||||||
def draw_protocol_graph(protocol_graph: SimpleGraph, output_path: str):
|
|
||||||
"""
|
|
||||||
(辅助功能) 使用 networkx 和 matplotlib 绘制协议工作流图,用于可视化。
|
|
||||||
"""
|
|
||||||
if not protocol_graph:
|
|
||||||
print("Cannot draw graph: Graph object is empty.")
|
|
||||||
return
|
|
||||||
|
|
||||||
G = nx.DiGraph()
|
|
||||||
|
|
||||||
for node_id, attrs in protocol_graph.nodes.items():
|
|
||||||
label = attrs.get("description", attrs.get("template", node_id[:8]))
|
|
||||||
G.add_node(node_id, label=label, **attrs)
|
|
||||||
|
|
||||||
for edge in protocol_graph.edges:
|
|
||||||
G.add_edge(edge["source"], edge["target"])
|
|
||||||
|
|
||||||
plt.figure(figsize=(20, 15))
|
|
||||||
try:
|
|
||||||
pos = nx.nx_agraph.graphviz_layout(G, prog="dot")
|
|
||||||
except Exception:
|
|
||||||
pos = nx.shell_layout(G) # Fallback layout
|
|
||||||
|
|
||||||
node_labels = {node: data["label"] for node, data in G.nodes(data=True)}
|
|
||||||
nx.draw(
|
|
||||||
G,
|
|
||||||
pos,
|
|
||||||
with_labels=False,
|
|
||||||
node_size=2500,
|
|
||||||
node_color="skyblue",
|
|
||||||
node_shape="o",
|
|
||||||
edge_color="gray",
|
|
||||||
width=1.5,
|
|
||||||
arrowsize=15,
|
|
||||||
)
|
|
||||||
nx.draw_networkx_labels(G, pos, labels=node_labels, font_size=8, font_weight="bold")
|
|
||||||
|
|
||||||
plt.title("Chemical Protocol Workflow Graph", size=15)
|
|
||||||
plt.savefig(output_path, dpi=300, bbox_inches="tight")
|
|
||||||
plt.close()
|
|
||||||
print(f" - Visualization saved to '{output_path}'")
|
|
||||||
|
|
||||||
|
|
||||||
from networkx.drawing.nx_agraph import to_agraph
|
|
||||||
import re
|
|
||||||
|
|
||||||
COMPASS = {"n","e","s","w","ne","nw","se","sw","c"}
|
|
||||||
|
|
||||||
def _is_compass(port: str) -> bool:
|
|
||||||
return isinstance(port, str) and port.lower() in COMPASS
|
|
||||||
|
|
||||||
def draw_protocol_graph_with_ports(protocol_graph, output_path: str, rankdir: str = "LR"):
|
|
||||||
"""
|
|
||||||
使用 Graphviz 端口语法绘制协议工作流图。
|
|
||||||
- 若边上的 source_port/target_port 是 compass(n/e/s/w/...),直接用 compass。
|
|
||||||
- 否则自动为节点创建 record 形状并定义命名端口 <portname>。
|
|
||||||
最终由 PyGraphviz 渲染并输出到 output_path(后缀决定格式,如 .png/.svg/.pdf)。
|
|
||||||
"""
|
|
||||||
if not protocol_graph:
|
|
||||||
print("Cannot draw graph: Graph object is empty.")
|
|
||||||
return
|
|
||||||
|
|
||||||
# 1) 先用 networkx 搭建有向图,保留端口属性
|
|
||||||
G = nx.DiGraph()
|
|
||||||
for node_id, attrs in protocol_graph.nodes.items():
|
|
||||||
label = attrs.get("description", attrs.get("template", node_id[:8]))
|
|
||||||
# 保留一个干净的“中心标签”,用于放在 record 的中间槽
|
|
||||||
G.add_node(node_id, _core_label=str(label), **{k:v for k,v in attrs.items() if k not in ("label",)})
|
|
||||||
|
|
||||||
edges_data = []
|
|
||||||
in_ports_by_node = {} # 收集命名输入端口
|
|
||||||
out_ports_by_node = {} # 收集命名输出端口
|
|
||||||
|
|
||||||
for edge in protocol_graph.edges:
|
|
||||||
u = edge["source"]
|
|
||||||
v = edge["target"]
|
|
||||||
sp = edge.get("source_port")
|
|
||||||
tp = edge.get("target_port")
|
|
||||||
|
|
||||||
# 记录到图里(保留原始端口信息)
|
|
||||||
G.add_edge(u, v, source_port=sp, target_port=tp)
|
|
||||||
edges_data.append((u, v, sp, tp))
|
|
||||||
|
|
||||||
# 如果不是 compass,就按“命名端口”先归类,等会儿给节点造 record
|
|
||||||
if sp and not _is_compass(sp):
|
|
||||||
out_ports_by_node.setdefault(u, set()).add(str(sp))
|
|
||||||
if tp and not _is_compass(tp):
|
|
||||||
in_ports_by_node.setdefault(v, set()).add(str(tp))
|
|
||||||
|
|
||||||
# 2) 转为 AGraph,使用 Graphviz 渲染
|
|
||||||
A = to_agraph(G)
|
|
||||||
A.graph_attr.update(rankdir=rankdir, splines="true", concentrate="false", fontsize="10")
|
|
||||||
A.node_attr.update(shape="box", style="rounded,filled", fillcolor="lightyellow", color="#999999", fontname="Helvetica")
|
|
||||||
A.edge_attr.update(arrowsize="0.8", color="#666666")
|
|
||||||
|
|
||||||
# 3) 为需要命名端口的节点设置 record 形状与 label
|
|
||||||
# 左列 = 输入端口;中间 = 核心标签;右列 = 输出端口
|
|
||||||
for n in A.nodes():
|
|
||||||
node = A.get_node(n)
|
|
||||||
core = G.nodes[n].get("_core_label", n)
|
|
||||||
|
|
||||||
in_ports = sorted(in_ports_by_node.get(n, []))
|
|
||||||
out_ports = sorted(out_ports_by_node.get(n, []))
|
|
||||||
|
|
||||||
# 如果该节点涉及命名端口,则用 record;否则保留原 box
|
|
||||||
if in_ports or out_ports:
|
|
||||||
def port_fields(ports):
|
|
||||||
if not ports:
|
|
||||||
return " " # 必须留一个空槽占位
|
|
||||||
# 每个端口一个小格子,<p> name
|
|
||||||
return "|".join(f"<{re.sub(r'[^A-Za-z0-9_:.|-]', '_', p)}> {p}" for p in ports)
|
|
||||||
|
|
||||||
left = port_fields(in_ports)
|
|
||||||
right = port_fields(out_ports)
|
|
||||||
|
|
||||||
# 三栏:左(入) | 中(节点名) | 右(出)
|
|
||||||
record_label = f"{{ {left} | {core} | {right} }}"
|
|
||||||
node.attr.update(shape="record", label=record_label)
|
|
||||||
else:
|
|
||||||
# 没有命名端口:普通盒子,显示核心标签
|
|
||||||
node.attr.update(label=str(core))
|
|
||||||
|
|
||||||
# 4) 给边设置 headport / tailport
|
|
||||||
# - 若端口为 compass:直接用 compass(e.g., headport="e")
|
|
||||||
# - 若端口为命名端口:使用在 record 中定义的 <port> 名(同名即可)
|
|
||||||
for (u, v, sp, tp) in edges_data:
|
|
||||||
e = A.get_edge(u, v)
|
|
||||||
|
|
||||||
# Graphviz 属性:tail 是源,head 是目标
|
|
||||||
if sp:
|
|
||||||
if _is_compass(sp):
|
|
||||||
e.attr["tailport"] = sp.lower()
|
|
||||||
else:
|
|
||||||
# 与 record label 中 <port> 名一致;特殊字符已在 label 中做了清洗
|
|
||||||
e.attr["tailport"] = re.sub(r'[^A-Za-z0-9_:.|-]', '_', str(sp))
|
|
||||||
|
|
||||||
if tp:
|
|
||||||
if _is_compass(tp):
|
|
||||||
e.attr["headport"] = tp.lower()
|
|
||||||
else:
|
|
||||||
e.attr["headport"] = re.sub(r'[^A-Za-z0-9_:.|-]', '_', str(tp))
|
|
||||||
|
|
||||||
# 可选:若想让边更贴边缘,可设置 constraint/spline 等
|
|
||||||
# e.attr["arrowhead"] = "vee"
|
|
||||||
|
|
||||||
# 5) 输出
|
|
||||||
A.draw(output_path, prog="dot")
|
|
||||||
print(f" - Port-aware workflow rendered to '{output_path}'")
|
|
||||||
|
|
||||||
|
|
||||||
def flatten_xdl_procedure(procedure_elem: ET.Element) -> List[ET.Element]:
|
|
||||||
"""展平嵌套的XDL程序结构"""
|
|
||||||
flattened_operations = []
|
|
||||||
TEMP_UNSUPPORTED_PROTOCOL = ["Purge", "Wait", "Stir", "ResetHandling"]
|
|
||||||
|
|
||||||
def extract_operations(element: ET.Element):
|
|
||||||
if element.tag not in ["Prep", "Reaction", "Workup", "Purification", "Procedure"]:
|
|
||||||
if element.tag not in TEMP_UNSUPPORTED_PROTOCOL:
|
|
||||||
flattened_operations.append(element)
|
|
||||||
|
|
||||||
for child in element:
|
|
||||||
extract_operations(child)
|
|
||||||
|
|
||||||
for child in procedure_elem:
|
|
||||||
extract_operations(child)
|
|
||||||
|
|
||||||
return flattened_operations
|
|
||||||
|
|
||||||
|
|
||||||
def parse_xdl_content(xdl_content: str) -> tuple:
|
|
||||||
"""解析XDL内容"""
|
|
||||||
try:
|
|
||||||
xdl_content_cleaned = "".join(c for c in xdl_content if c.isprintable())
|
|
||||||
root = ET.fromstring(xdl_content_cleaned)
|
|
||||||
|
|
||||||
synthesis_elem = root.find("Synthesis")
|
|
||||||
if synthesis_elem is None:
|
|
||||||
return None, None, None
|
|
||||||
|
|
||||||
# 解析硬件组件
|
|
||||||
hardware_elem = synthesis_elem.find("Hardware")
|
|
||||||
hardware = []
|
|
||||||
if hardware_elem is not None:
|
|
||||||
hardware = [{"id": c.get("id"), "type": c.get("type")} for c in hardware_elem.findall("Component")]
|
|
||||||
|
|
||||||
# 解析试剂
|
|
||||||
reagents_elem = synthesis_elem.find("Reagents")
|
|
||||||
reagents = []
|
|
||||||
if reagents_elem is not None:
|
|
||||||
reagents = [{"name": r.get("name"), "role": r.get("role", "")} for r in reagents_elem.findall("Reagent")]
|
|
||||||
|
|
||||||
# 解析程序
|
|
||||||
procedure_elem = synthesis_elem.find("Procedure")
|
|
||||||
if procedure_elem is None:
|
|
||||||
return None, None, None
|
|
||||||
|
|
||||||
flattened_operations = flatten_xdl_procedure(procedure_elem)
|
|
||||||
return hardware, reagents, flattened_operations
|
|
||||||
|
|
||||||
except ET.ParseError as e:
|
|
||||||
raise ValueError(f"Invalid XDL format: {e}")
|
|
||||||
|
|
||||||
|
|
||||||
def convert_xdl_to_dict(xdl_content: str) -> Dict[str, Any]:
|
|
||||||
"""
|
|
||||||
将XDL XML格式转换为标准的字典格式
|
|
||||||
|
|
||||||
Args:
|
|
||||||
xdl_content: XDL XML内容
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
转换结果,包含步骤和器材信息
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
hardware, reagents, flattened_operations = parse_xdl_content(xdl_content)
|
|
||||||
if hardware is None:
|
|
||||||
return {"error": "Failed to parse XDL content", "success": False}
|
|
||||||
|
|
||||||
# 将XDL元素转换为字典格式
|
|
||||||
steps_data = []
|
|
||||||
for elem in flattened_operations:
|
|
||||||
# 转换参数类型
|
|
||||||
parameters = {}
|
|
||||||
for key, val in elem.attrib.items():
|
|
||||||
converted_val = convert_to_type(val)
|
|
||||||
if converted_val is not None:
|
|
||||||
parameters[key] = converted_val
|
|
||||||
|
|
||||||
step_dict = {
|
|
||||||
"operation": elem.tag,
|
|
||||||
"parameters": parameters,
|
|
||||||
"description": elem.get("purpose", f"Operation: {elem.tag}"),
|
|
||||||
}
|
|
||||||
steps_data.append(step_dict)
|
|
||||||
|
|
||||||
# 合并硬件和试剂为统一的labware_info格式
|
|
||||||
labware_data = []
|
|
||||||
labware_data.extend({"id": hw["id"], "type": "hardware", **hw} for hw in hardware)
|
|
||||||
labware_data.extend({"name": reagent["name"], "type": "reagent", **reagent} for reagent in reagents)
|
|
||||||
|
|
||||||
return {
|
|
||||||
"success": True,
|
|
||||||
"steps": steps_data,
|
|
||||||
"labware": labware_data,
|
|
||||||
"message": f"Successfully converted XDL to dict format. Found {len(steps_data)} steps and {len(labware_data)} labware items.",
|
|
||||||
}
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
error_msg = f"XDL conversion failed: {str(e)}"
|
|
||||||
logger.error(error_msg)
|
|
||||||
return {"error": error_msg, "success": False}
|
|
||||||
|
|
||||||
|
|
||||||
def create_workflow(
|
def create_workflow(
|
||||||
|
|||||||
2
setup.py
2
setup.py
@@ -4,7 +4,7 @@ package_name = 'unilabos'
|
|||||||
|
|
||||||
setup(
|
setup(
|
||||||
name=package_name,
|
name=package_name,
|
||||||
version='0.10.17',
|
version='0.10.15',
|
||||||
packages=find_packages(),
|
packages=find_packages(),
|
||||||
include_package_data=True,
|
include_package_data=True,
|
||||||
install_requires=['setuptools'],
|
install_requires=['setuptools'],
|
||||||
|
|||||||
@@ -1,213 +0,0 @@
|
|||||||
{
|
|
||||||
"workflow": [
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines",
|
|
||||||
"targets": "Liquid_1",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines",
|
|
||||||
"targets": "Liquid_2",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines",
|
|
||||||
"targets": "Liquid_3",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_2",
|
|
||||||
"targets": "Liquid_4",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_2",
|
|
||||||
"targets": "Liquid_5",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_2",
|
|
||||||
"targets": "Liquid_6",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_3",
|
|
||||||
"targets": "dest_set",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_3",
|
|
||||||
"targets": "dest_set_2",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"action": "transfer_liquid",
|
|
||||||
"action_args": {
|
|
||||||
"sources": "cell_lines_3",
|
|
||||||
"targets": "dest_set_3",
|
|
||||||
"asp_vol": 100.0,
|
|
||||||
"dis_vol": 74.75,
|
|
||||||
"asp_flow_rate": 94.0,
|
|
||||||
"dis_flow_rate": 95.5
|
|
||||||
}
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"reagent": {
|
|
||||||
"Liquid_1": {
|
|
||||||
"slot": 1,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A7",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 1"
|
|
||||||
},
|
|
||||||
"Liquid_4": {
|
|
||||||
"slot": 1,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A7",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 1"
|
|
||||||
},
|
|
||||||
"dest_set": {
|
|
||||||
"slot": 1,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A7",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 1"
|
|
||||||
},
|
|
||||||
"Liquid_2": {
|
|
||||||
"slot": 2,
|
|
||||||
"well": [
|
|
||||||
"A3",
|
|
||||||
"A5",
|
|
||||||
"A8"
|
|
||||||
],
|
|
||||||
"labware": "rep 2"
|
|
||||||
},
|
|
||||||
"Liquid_5": {
|
|
||||||
"slot": 2,
|
|
||||||
"well": [
|
|
||||||
"A3",
|
|
||||||
"A5",
|
|
||||||
"A8"
|
|
||||||
],
|
|
||||||
"labware": "rep 2"
|
|
||||||
},
|
|
||||||
"dest_set_2": {
|
|
||||||
"slot": 2,
|
|
||||||
"well": [
|
|
||||||
"A3",
|
|
||||||
"A5",
|
|
||||||
"A8"
|
|
||||||
],
|
|
||||||
"labware": "rep 2"
|
|
||||||
},
|
|
||||||
"Liquid_3": {
|
|
||||||
"slot": 3,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A6",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 3"
|
|
||||||
},
|
|
||||||
"Liquid_6": {
|
|
||||||
"slot": 3,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A6",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 3"
|
|
||||||
},
|
|
||||||
"dest_set_3": {
|
|
||||||
"slot": 3,
|
|
||||||
"well": [
|
|
||||||
"A4",
|
|
||||||
"A6",
|
|
||||||
"A10"
|
|
||||||
],
|
|
||||||
"labware": "rep 3"
|
|
||||||
},
|
|
||||||
"cell_lines": {
|
|
||||||
"slot": 4,
|
|
||||||
"well": [
|
|
||||||
"A1",
|
|
||||||
"A3",
|
|
||||||
"A5"
|
|
||||||
],
|
|
||||||
"labware": "DRUG + YOYO-MEDIA"
|
|
||||||
},
|
|
||||||
"cell_lines_2": {
|
|
||||||
"slot": 4,
|
|
||||||
"well": [
|
|
||||||
"A1",
|
|
||||||
"A3",
|
|
||||||
"A5"
|
|
||||||
],
|
|
||||||
"labware": "DRUG + YOYO-MEDIA"
|
|
||||||
},
|
|
||||||
"cell_lines_3": {
|
|
||||||
"slot": 4,
|
|
||||||
"well": [
|
|
||||||
"A1",
|
|
||||||
"A3",
|
|
||||||
"A5"
|
|
||||||
],
|
|
||||||
"labware": "DRUG + YOYO-MEDIA"
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1 +1 @@
|
|||||||
__version__ = "0.10.17"
|
__version__ = "0.10.15"
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
"""Entry point for `python -m unilabos`."""
|
|
||||||
|
|
||||||
from unilabos.app.main import main
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
||||||
@@ -7,6 +7,7 @@ import sys
|
|||||||
import threading
|
import threading
|
||||||
import time
|
import time
|
||||||
from typing import Dict, Any, List
|
from typing import Dict, Any, List
|
||||||
|
|
||||||
import networkx as nx
|
import networkx as nx
|
||||||
import yaml
|
import yaml
|
||||||
|
|
||||||
@@ -16,9 +17,9 @@ unilabos_dir = os.path.dirname(os.path.dirname(current_dir))
|
|||||||
if unilabos_dir not in sys.path:
|
if unilabos_dir not in sys.path:
|
||||||
sys.path.append(unilabos_dir)
|
sys.path.append(unilabos_dir)
|
||||||
|
|
||||||
from unilabos.app.utils import cleanup_for_restart
|
|
||||||
from unilabos.utils.banner_print import print_status, print_unilab_banner
|
from unilabos.utils.banner_print import print_status, print_unilab_banner
|
||||||
from unilabos.config.config import load_config, BasicConfig, HTTPConfig
|
from unilabos.config.config import load_config, BasicConfig, HTTPConfig
|
||||||
|
from unilabos.app.utils import cleanup_for_restart
|
||||||
|
|
||||||
# Global restart flags (used by ws_client and web/server)
|
# Global restart flags (used by ws_client and web/server)
|
||||||
_restart_requested: bool = False
|
_restart_requested: bool = False
|
||||||
@@ -216,10 +217,7 @@ def main():
|
|||||||
args_dict = vars(args)
|
args_dict = vars(args)
|
||||||
|
|
||||||
# 环境检查 - 检查并自动安装必需的包 (可选)
|
# 环境检查 - 检查并自动安装必需的包 (可选)
|
||||||
skip_env_check = args_dict.get("skip_env_check", False)
|
if not args_dict.get("skip_env_check", False):
|
||||||
check_mode = args_dict.get("check_mode", False)
|
|
||||||
|
|
||||||
if not skip_env_check:
|
|
||||||
from unilabos.utils.environment_check import check_environment
|
from unilabos.utils.environment_check import check_environment
|
||||||
|
|
||||||
if not check_environment(auto_install=True):
|
if not check_environment(auto_install=True):
|
||||||
@@ -230,21 +228,7 @@ def main():
|
|||||||
|
|
||||||
# 加载配置文件,优先加载config,然后从env读取
|
# 加载配置文件,优先加载config,然后从env读取
|
||||||
config_path = args_dict.get("config")
|
config_path = args_dict.get("config")
|
||||||
|
if os.getcwd().endswith("unilabos_data"):
|
||||||
if check_mode:
|
|
||||||
args_dict["working_dir"] = os.path.abspath(os.getcwd())
|
|
||||||
# 当 skip_env_check 时,默认使用当前目录作为 working_dir
|
|
||||||
if skip_env_check and not args_dict.get("working_dir") and not config_path:
|
|
||||||
working_dir = os.path.abspath(os.getcwd())
|
|
||||||
print_status(f"跳过环境检查模式:使用当前目录作为工作目录 {working_dir}", "info")
|
|
||||||
# 检查当前目录是否有 local_config.py
|
|
||||||
local_config_in_cwd = os.path.join(working_dir, "local_config.py")
|
|
||||||
if os.path.exists(local_config_in_cwd):
|
|
||||||
config_path = local_config_in_cwd
|
|
||||||
print_status(f"发现本地配置文件: {config_path}", "info")
|
|
||||||
else:
|
|
||||||
print_status(f"未指定config路径,可通过 --config 传入 local_config.py 文件路径", "info")
|
|
||||||
elif os.getcwd().endswith("unilabos_data"):
|
|
||||||
working_dir = os.path.abspath(os.getcwd())
|
working_dir = os.path.abspath(os.getcwd())
|
||||||
else:
|
else:
|
||||||
working_dir = os.path.abspath(os.path.join(os.getcwd(), "unilabos_data"))
|
working_dir = os.path.abspath(os.path.join(os.getcwd(), "unilabos_data"))
|
||||||
@@ -263,7 +247,7 @@ def main():
|
|||||||
working_dir = os.path.dirname(config_path)
|
working_dir = os.path.dirname(config_path)
|
||||||
elif os.path.exists(working_dir) and os.path.exists(os.path.join(working_dir, "local_config.py")):
|
elif os.path.exists(working_dir) and os.path.exists(os.path.join(working_dir, "local_config.py")):
|
||||||
config_path = os.path.join(working_dir, "local_config.py")
|
config_path = os.path.join(working_dir, "local_config.py")
|
||||||
elif not skip_env_check and not config_path and (
|
elif not config_path and (
|
||||||
not os.path.exists(working_dir) or not os.path.exists(os.path.join(working_dir, "local_config.py"))
|
not os.path.exists(working_dir) or not os.path.exists(os.path.join(working_dir, "local_config.py"))
|
||||||
):
|
):
|
||||||
print_status(f"未指定config路径,可通过 --config 传入 local_config.py 文件路径", "info")
|
print_status(f"未指定config路径,可通过 --config 传入 local_config.py 文件路径", "info")
|
||||||
@@ -277,10 +261,8 @@ def main():
|
|||||||
print_status(f"已创建 local_config.py 路径: {config_path}", "info")
|
print_status(f"已创建 local_config.py 路径: {config_path}", "info")
|
||||||
else:
|
else:
|
||||||
os._exit(1)
|
os._exit(1)
|
||||||
|
# 加载配置文件
|
||||||
# 加载配置文件 (check_mode 跳过)
|
|
||||||
print_status(f"当前工作目录为 {working_dir}", "info")
|
print_status(f"当前工作目录为 {working_dir}", "info")
|
||||||
if not check_mode:
|
|
||||||
load_config_from_file(config_path)
|
load_config_from_file(config_path)
|
||||||
|
|
||||||
# 根据配置重新设置日志级别
|
# 根据配置重新设置日志级别
|
||||||
@@ -337,7 +319,12 @@ def main():
|
|||||||
machine_name = "".join([c if c.isalnum() or c == "_" else "_" for c in machine_name])
|
machine_name = "".join([c if c.isalnum() or c == "_" else "_" for c in machine_name])
|
||||||
BasicConfig.machine_name = machine_name
|
BasicConfig.machine_name = machine_name
|
||||||
BasicConfig.vis_2d_enable = args_dict["2d_vis"]
|
BasicConfig.vis_2d_enable = args_dict["2d_vis"]
|
||||||
|
|
||||||
|
# Check mode 处理
|
||||||
|
check_mode = args_dict.get("check_mode", False)
|
||||||
BasicConfig.check_mode = check_mode
|
BasicConfig.check_mode = check_mode
|
||||||
|
if check_mode:
|
||||||
|
print_status("Check mode 启用,将进行 complete_registry 检查", "info")
|
||||||
|
|
||||||
from unilabos.resources.graphio import (
|
from unilabos.resources.graphio import (
|
||||||
read_node_link_json,
|
read_node_link_json,
|
||||||
|
|||||||
@@ -4,40 +4,8 @@ UniLabOS 应用工具函数
|
|||||||
提供清理、重启等工具函数
|
提供清理、重启等工具函数
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import glob
|
|
||||||
import os
|
|
||||||
import shutil
|
|
||||||
import sys
|
|
||||||
|
|
||||||
|
|
||||||
def patch_rclpy_dll_windows():
|
|
||||||
"""在 Windows + conda 环境下为 rclpy 打 DLL 加载补丁"""
|
|
||||||
if sys.platform != "win32" or not os.environ.get("CONDA_PREFIX"):
|
|
||||||
return
|
|
||||||
try:
|
|
||||||
import rclpy
|
|
||||||
|
|
||||||
return
|
|
||||||
except ImportError as e:
|
|
||||||
if not str(e).startswith("DLL load failed"):
|
|
||||||
return
|
|
||||||
cp = os.environ["CONDA_PREFIX"]
|
|
||||||
impl = os.path.join(cp, "Lib", "site-packages", "rclpy", "impl", "implementation_singleton.py")
|
|
||||||
pyd = glob.glob(os.path.join(cp, "Lib", "site-packages", "rclpy", "_rclpy_pybind11*.pyd"))
|
|
||||||
if not os.path.exists(impl) or not pyd:
|
|
||||||
return
|
|
||||||
with open(impl, "r", encoding="utf-8") as f:
|
|
||||||
content = f.read()
|
|
||||||
lib_bin = os.path.join(cp, "Library", "bin").replace("\\", "/")
|
|
||||||
patch = f'# UniLabOS DLL Patch\nimport os,ctypes\nos.add_dll_directory("{lib_bin}") if hasattr(os,"add_dll_directory") else None\ntry: ctypes.CDLL("{pyd[0].replace(chr(92),"/")}")\nexcept: pass\n# End Patch\n'
|
|
||||||
shutil.copy2(impl, impl + ".bak")
|
|
||||||
with open(impl, "w", encoding="utf-8") as f:
|
|
||||||
f.write(patch + content)
|
|
||||||
|
|
||||||
|
|
||||||
patch_rclpy_dll_windows()
|
|
||||||
|
|
||||||
import gc
|
import gc
|
||||||
|
import os
|
||||||
import threading
|
import threading
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
|||||||
@@ -359,7 +359,9 @@ class HTTPClient:
|
|||||||
Returns:
|
Returns:
|
||||||
Dict: API响应数据,包含 code 和 data (uuid, name)
|
Dict: API响应数据,包含 code 和 data (uuid, name)
|
||||||
"""
|
"""
|
||||||
|
# target_lab_uuid 暂时使用默认值,后续由后端根据 ak/sk 获取
|
||||||
payload = {
|
payload = {
|
||||||
|
"target_lab_uuid": "28c38bb0-63f6-4352-b0d8-b5b8eb1766d5",
|
||||||
"name": name,
|
"name": name,
|
||||||
"data": {
|
"data": {
|
||||||
"workflow_uuid": workflow_uuid,
|
"workflow_uuid": workflow_uuid,
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ from typing import Optional, Dict, Any, List
|
|||||||
from urllib.parse import urlparse
|
from urllib.parse import urlparse
|
||||||
from enum import Enum
|
from enum import Enum
|
||||||
|
|
||||||
from typing_extensions import TypedDict
|
from jedi.inference.gradual.typing import TypedDict
|
||||||
|
|
||||||
from unilabos.app.model import JobAddReq
|
from unilabos.app.model import JobAddReq
|
||||||
from unilabos.ros.nodes.presets.host_node import HostNode
|
from unilabos.ros.nodes.presets.host_node import HostNode
|
||||||
@@ -439,7 +439,7 @@ class MessageProcessor:
|
|||||||
self.connected = True
|
self.connected = True
|
||||||
self.reconnect_count = 0
|
self.reconnect_count = 0
|
||||||
|
|
||||||
logger.info(f"[MessageProcessor] Connected to {self.websocket_url}")
|
logger.trace(f"[MessageProcessor] Connected to {self.websocket_url}")
|
||||||
|
|
||||||
# 启动发送协程
|
# 启动发送协程
|
||||||
send_task = asyncio.create_task(self._send_handler())
|
send_task = asyncio.create_task(self._send_handler())
|
||||||
@@ -495,12 +495,8 @@ class MessageProcessor:
|
|||||||
await self._process_message(message_type, message_data)
|
await self._process_message(message_type, message_data)
|
||||||
else:
|
else:
|
||||||
if message_type.endswith("_material"):
|
if message_type.endswith("_material"):
|
||||||
logger.trace(
|
logger.trace(f"[MessageProcessor] 收到一条归属 {data.get('edge_session')} 的旧消息:{data}")
|
||||||
f"[MessageProcessor] 收到一条归属 {data.get('edge_session')} 的旧消息:{data}"
|
logger.debug(f"[MessageProcessor] 跳过了一条归属 {data.get('edge_session')} 的旧消息: {data.get('action')}")
|
||||||
)
|
|
||||||
logger.debug(
|
|
||||||
f"[MessageProcessor] 跳过了一条归属 {data.get('edge_session')} 的旧消息: {data.get('action')}"
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
await self._process_message(message_type, message_data)
|
await self._process_message(message_type, message_data)
|
||||||
except json.JSONDecodeError:
|
except json.JSONDecodeError:
|
||||||
@@ -517,7 +513,7 @@ class MessageProcessor:
|
|||||||
|
|
||||||
async def _send_handler(self):
|
async def _send_handler(self):
|
||||||
"""处理发送队列中的消息"""
|
"""处理发送队列中的消息"""
|
||||||
logger.debug("[MessageProcessor] Send handler started")
|
logger.trace("[MessageProcessor] Send handler started")
|
||||||
|
|
||||||
try:
|
try:
|
||||||
while self.connected and self.websocket:
|
while self.connected and self.websocket:
|
||||||
@@ -852,7 +848,9 @@ class MessageProcessor:
|
|||||||
device_action_groups[key_add] = []
|
device_action_groups[key_add] = []
|
||||||
device_action_groups[key_add].append(item["uuid"])
|
device_action_groups[key_add].append(item["uuid"])
|
||||||
|
|
||||||
logger.info(f"[资源同步] 跨站Transfer: {item['uuid'][:8]} from {device_old_id} to {device_id}")
|
logger.info(
|
||||||
|
f"[资源同步] 跨站Transfer: {item['uuid'][:8]} from {device_old_id} to {device_id}"
|
||||||
|
)
|
||||||
else:
|
else:
|
||||||
# 正常update
|
# 正常update
|
||||||
key = (device_id, "update")
|
key = (device_id, "update")
|
||||||
@@ -866,9 +864,7 @@ class MessageProcessor:
|
|||||||
device_action_groups[key] = []
|
device_action_groups[key] = []
|
||||||
device_action_groups[key].append(item["uuid"])
|
device_action_groups[key].append(item["uuid"])
|
||||||
|
|
||||||
logger.trace(
|
logger.trace(f"[资源同步] 动作 {action} 分组数量: {len(device_action_groups)}, 总数量: {len(resource_uuid_list)}")
|
||||||
f"[资源同步] 动作 {action} 分组数量: {len(device_action_groups)}, 总数量: {len(resource_uuid_list)}"
|
|
||||||
)
|
|
||||||
|
|
||||||
# 为每个(device_id, action)创建独立的更新线程
|
# 为每个(device_id, action)创建独立的更新线程
|
||||||
for (device_id, actual_action), items in device_action_groups.items():
|
for (device_id, actual_action), items in device_action_groups.items():
|
||||||
@@ -916,13 +912,13 @@ class MessageProcessor:
|
|||||||
|
|
||||||
# 发送确认消息
|
# 发送确认消息
|
||||||
if self.websocket_client:
|
if self.websocket_client:
|
||||||
await self.websocket_client.send_message(
|
await self.websocket_client.send_message({
|
||||||
{"action": "restart_acknowledged", "data": {"reason": reason, "delay": delay}}
|
"action": "restart_acknowledged",
|
||||||
)
|
"data": {"reason": reason, "delay": delay}
|
||||||
|
})
|
||||||
|
|
||||||
# 设置全局重启标志
|
# 设置全局重启标志
|
||||||
import unilabos.app.main as main_module
|
import unilabos.app.main as main_module
|
||||||
|
|
||||||
main_module._restart_requested = True
|
main_module._restart_requested = True
|
||||||
main_module._restart_reason = reason
|
main_module._restart_reason = reason
|
||||||
|
|
||||||
@@ -932,12 +928,10 @@ class MessageProcessor:
|
|||||||
# 在新线程中执行清理,避免阻塞当前事件循环
|
# 在新线程中执行清理,避免阻塞当前事件循环
|
||||||
def do_cleanup():
|
def do_cleanup():
|
||||||
import time
|
import time
|
||||||
|
|
||||||
time.sleep(0.5) # 给当前消息处理完成的时间
|
time.sleep(0.5) # 给当前消息处理完成的时间
|
||||||
logger.info(f"[MessageProcessor] Starting cleanup for restart, reason: {reason}")
|
logger.info(f"[MessageProcessor] Starting cleanup for restart, reason: {reason}")
|
||||||
try:
|
try:
|
||||||
from unilabos.app.utils import cleanup_for_restart
|
from unilabos.app.utils import cleanup_for_restart
|
||||||
|
|
||||||
if cleanup_for_restart():
|
if cleanup_for_restart():
|
||||||
logger.info("[MessageProcessor] Cleanup successful, main() will restart")
|
logger.info("[MessageProcessor] Cleanup successful, main() will restart")
|
||||||
else:
|
else:
|
||||||
@@ -1026,7 +1020,7 @@ class QueueProcessor:
|
|||||||
|
|
||||||
def _run(self):
|
def _run(self):
|
||||||
"""运行队列处理主循环"""
|
"""运行队列处理主循环"""
|
||||||
logger.debug("[QueueProcessor] Queue processor started")
|
logger.trace("[QueueProcessor] Queue processor started")
|
||||||
|
|
||||||
while self.is_running:
|
while self.is_running:
|
||||||
try:
|
try:
|
||||||
@@ -1236,7 +1230,6 @@ class WebSocketClient(BaseCommunicationClient):
|
|||||||
else:
|
else:
|
||||||
url = f"{scheme}://{parsed.netloc}/api/v1/ws/schedule"
|
url = f"{scheme}://{parsed.netloc}/api/v1/ws/schedule"
|
||||||
|
|
||||||
logger.debug(f"[WebSocketClient] URL: {url}")
|
|
||||||
return url
|
return url
|
||||||
|
|
||||||
def start(self) -> None:
|
def start(self) -> None:
|
||||||
@@ -1249,13 +1242,11 @@ class WebSocketClient(BaseCommunicationClient):
|
|||||||
logger.error("[WebSocketClient] WebSocket URL not configured")
|
logger.error("[WebSocketClient] WebSocket URL not configured")
|
||||||
return
|
return
|
||||||
|
|
||||||
logger.info(f"[WebSocketClient] Starting connection to {self.websocket_url}")
|
|
||||||
|
|
||||||
# 启动两个核心线程
|
# 启动两个核心线程
|
||||||
self.message_processor.start()
|
self.message_processor.start()
|
||||||
self.queue_processor.start()
|
self.queue_processor.start()
|
||||||
|
|
||||||
logger.info("[WebSocketClient] All threads started")
|
logger.trace("[WebSocketClient] All threads started")
|
||||||
|
|
||||||
def stop(self) -> None:
|
def stop(self) -> None:
|
||||||
"""停止WebSocket客户端"""
|
"""停止WebSocket客户端"""
|
||||||
@@ -1391,9 +1382,7 @@ class WebSocketClient(BaseCommunicationClient):
|
|||||||
if host_node:
|
if host_node:
|
||||||
# 获取设备信息
|
# 获取设备信息
|
||||||
for device_id, namespace in host_node.devices_names.items():
|
for device_id, namespace in host_node.devices_names.items():
|
||||||
device_key = (
|
device_key = f"{namespace}/{device_id}" if namespace.startswith("/") else f"/{namespace}/{device_id}"
|
||||||
f"{namespace}/{device_id}" if namespace.startswith("/") else f"/{namespace}/{device_id}"
|
|
||||||
)
|
|
||||||
is_online = device_key in host_node._online_devices
|
is_online = device_key in host_node._online_devices
|
||||||
|
|
||||||
# 获取设备的动作信息
|
# 获取设备的动作信息
|
||||||
@@ -1407,16 +1396,14 @@ class WebSocketClient(BaseCommunicationClient):
|
|||||||
"action_type": str(type(client).__name__),
|
"action_type": str(type(client).__name__),
|
||||||
}
|
}
|
||||||
|
|
||||||
devices.append(
|
devices.append({
|
||||||
{
|
|
||||||
"device_id": device_id,
|
"device_id": device_id,
|
||||||
"namespace": namespace,
|
"namespace": namespace,
|
||||||
"device_key": device_key,
|
"device_key": device_key,
|
||||||
"is_online": is_online,
|
"is_online": is_online,
|
||||||
"machine_name": host_node.device_machine_names.get(device_id, machine_name),
|
"machine_name": host_node.device_machine_names.get(device_id, machine_name),
|
||||||
"actions": actions,
|
"actions": actions,
|
||||||
}
|
})
|
||||||
)
|
|
||||||
|
|
||||||
logger.info(f"[WebSocketClient] Collected {len(devices)} devices for host_ready")
|
logger.info(f"[WebSocketClient] Collected {len(devices)} devices for host_ready")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
|||||||
@@ -28,40 +28,21 @@ from pylabrobot.resources import (
|
|||||||
Tip,
|
Tip,
|
||||||
)
|
)
|
||||||
|
|
||||||
from unilabos.registry.placeholder_type import ResourceSlot
|
|
||||||
from unilabos.ros.nodes.base_device_node import BaseROS2DeviceNode
|
from unilabos.ros.nodes.base_device_node import BaseROS2DeviceNode
|
||||||
from unilabos.resources.resource_tracker import ResourceTreeSet
|
|
||||||
|
|
||||||
|
|
||||||
class SimpleReturn(TypedDict):
|
class SimpleReturn(TypedDict):
|
||||||
samples: list
|
samples: list
|
||||||
volumes: list
|
volumes: list
|
||||||
|
|
||||||
|
|
||||||
class SetLiquidReturn(TypedDict):
|
|
||||||
wells: list
|
|
||||||
volumes: list
|
|
||||||
|
|
||||||
|
|
||||||
class SetLiquidFromPlateReturn(TypedDict):
|
|
||||||
plate: list
|
|
||||||
wells: list
|
|
||||||
volumes: list
|
|
||||||
|
|
||||||
|
|
||||||
class LiquidHandlerMiddleware(LiquidHandler):
|
class LiquidHandlerMiddleware(LiquidHandler):
|
||||||
def __init__(
|
def __init__(self, backend: LiquidHandlerBackend, deck: Deck, simulator: bool = False, channel_num: int = 8, **kwargs):
|
||||||
self, backend: LiquidHandlerBackend, deck: Deck, simulator: bool = False, channel_num: int = 8, **kwargs
|
|
||||||
):
|
|
||||||
self._simulator = simulator
|
self._simulator = simulator
|
||||||
self.channel_num = channel_num
|
self.channel_num = channel_num
|
||||||
self.pending_liquids_dict = {}
|
self.pending_liquids_dict = {}
|
||||||
joint_config = kwargs.get("joint_config", None)
|
joint_config = kwargs.get("joint_config", None)
|
||||||
if simulator:
|
if simulator:
|
||||||
if joint_config:
|
if joint_config:
|
||||||
self._simulate_backend = UniLiquidHandlerRvizBackend(
|
self._simulate_backend = UniLiquidHandlerRvizBackend(channel_num, kwargs["total_height"],
|
||||||
channel_num, kwargs["total_height"], joint_config=joint_config, lh_device_id=deck.name
|
joint_config=joint_config, lh_device_id=deck.name)
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
self._simulate_backend = LiquidHandlerChatterboxBackend(channel_num)
|
self._simulate_backend = LiquidHandlerChatterboxBackend(channel_num)
|
||||||
self._simulate_handler = LiquidHandlerAbstract(self._simulate_backend, deck, False)
|
self._simulate_handler = LiquidHandlerAbstract(self._simulate_backend, deck, False)
|
||||||
@@ -178,9 +159,7 @@ class LiquidHandlerMiddleware(LiquidHandler):
|
|||||||
if not offsets or (isinstance(offsets, list) and len(offsets) != len(use_channels)):
|
if not offsets or (isinstance(offsets, list) and len(offsets) != len(use_channels)):
|
||||||
offsets = [Coordinate.zero()] * len(use_channels)
|
offsets = [Coordinate.zero()] * len(use_channels)
|
||||||
if self._simulator:
|
if self._simulator:
|
||||||
return await self._simulate_handler.discard_tips(
|
return await self._simulate_handler.discard_tips(use_channels, allow_nonzero_volume, offsets, **backend_kwargs)
|
||||||
use_channels, allow_nonzero_volume, offsets, **backend_kwargs
|
|
||||||
)
|
|
||||||
await super().discard_tips(use_channels, allow_nonzero_volume, offsets, **backend_kwargs)
|
await super().discard_tips(use_channels, allow_nonzero_volume, offsets, **backend_kwargs)
|
||||||
self.pending_liquids_dict = {}
|
self.pending_liquids_dict = {}
|
||||||
return
|
return
|
||||||
@@ -201,6 +180,7 @@ class LiquidHandlerMiddleware(LiquidHandler):
|
|||||||
**backend_kwargs,
|
**backend_kwargs,
|
||||||
):
|
):
|
||||||
|
|
||||||
|
|
||||||
if self._simulator:
|
if self._simulator:
|
||||||
return await self._simulate_handler.aspirate(
|
return await self._simulate_handler.aspirate(
|
||||||
resources,
|
resources,
|
||||||
@@ -228,16 +208,15 @@ class LiquidHandlerMiddleware(LiquidHandler):
|
|||||||
res_samples = []
|
res_samples = []
|
||||||
res_volumes = []
|
res_volumes = []
|
||||||
for resource, volume, channel in zip(resources, vols, use_channels):
|
for resource, volume, channel in zip(resources, vols, use_channels):
|
||||||
res_samples.append(
|
res_samples.append({"name": resource.name, "sample_uuid": resource.unilabos_extra.get("sample_uuid", None)})
|
||||||
{"name": resource.name, "sample_uuid": resource.unilabos_extra.get("sample_uuid", None)}
|
|
||||||
)
|
|
||||||
res_volumes.append(volume)
|
res_volumes.append(volume)
|
||||||
self.pending_liquids_dict[channel] = {
|
self.pending_liquids_dict[channel] = {
|
||||||
"sample_uuid": resource.unilabos_extra.get("sample_uuid", None),
|
"sample_uuid": resource.unilabos_extra.get("sample_uuid", None),
|
||||||
"volume": volume,
|
"volume": volume
|
||||||
}
|
}
|
||||||
return SimpleReturn(samples=res_samples, volumes=res_volumes)
|
return SimpleReturn(samples=res_samples, volumes=res_volumes)
|
||||||
|
|
||||||
|
|
||||||
async def dispense(
|
async def dispense(
|
||||||
self,
|
self,
|
||||||
resources: Sequence[Container],
|
resources: Sequence[Container],
|
||||||
@@ -599,18 +578,10 @@ class LiquidHandlerMiddleware(LiquidHandler):
|
|||||||
|
|
||||||
class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
||||||
"""Extended LiquidHandler with additional operations."""
|
"""Extended LiquidHandler with additional operations."""
|
||||||
|
|
||||||
support_touch_tip = True
|
support_touch_tip = True
|
||||||
_ros_node: BaseROS2DeviceNode
|
_ros_node: BaseROS2DeviceNode
|
||||||
|
|
||||||
def __init__(
|
def __init__(self, backend: LiquidHandlerBackend, deck: Deck, simulator: bool=False, channel_num:int = 8, total_height:float = 310):
|
||||||
self,
|
|
||||||
backend: LiquidHandlerBackend,
|
|
||||||
deck: Deck,
|
|
||||||
simulator: bool = False,
|
|
||||||
channel_num: int = 8,
|
|
||||||
total_height: float = 310,
|
|
||||||
):
|
|
||||||
"""Initialize a LiquidHandler.
|
"""Initialize a LiquidHandler.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
@@ -634,7 +605,6 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
module_name = ".".join(components[:-1])
|
module_name = ".".join(components[:-1])
|
||||||
try:
|
try:
|
||||||
import importlib
|
import importlib
|
||||||
|
|
||||||
mod = importlib.import_module(module_name)
|
mod = importlib.import_module(module_name)
|
||||||
except ImportError:
|
except ImportError:
|
||||||
mod = None
|
mod = None
|
||||||
@@ -644,7 +614,6 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
# Try pylabrobot style import (if available)
|
# Try pylabrobot style import (if available)
|
||||||
try:
|
try:
|
||||||
import pylabrobot
|
import pylabrobot
|
||||||
|
|
||||||
backend_cls = getattr(pylabrobot, type_str, None)
|
backend_cls = getattr(pylabrobot, type_str, None)
|
||||||
except Exception:
|
except Exception:
|
||||||
backend_cls = None
|
backend_cls = None
|
||||||
@@ -662,56 +631,16 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
self._ros_node = ros_node
|
self._ros_node = ros_node
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
def set_liquid(cls, wells: list[Well], liquid_names: list[str], volumes: list[float]) -> SetLiquidReturn:
|
def set_liquid(cls, wells: list[Well], liquid_names: list[str], volumes: list[float]) -> SimpleReturn:
|
||||||
"""Set the liquid in a well.
|
"""Set the liquid in a well."""
|
||||||
|
res_samples = []
|
||||||
如果 liquid_names 和 volumes 为空,但 wells 不为空,直接返回 wells。
|
|
||||||
"""
|
|
||||||
res_volumes = []
|
res_volumes = []
|
||||||
# 如果 liquid_names 和 volumes 都为空,直接返回 wells
|
|
||||||
if not liquid_names and not volumes:
|
|
||||||
return SetLiquidReturn(
|
|
||||||
wells=ResourceTreeSet.from_plr_resources(wells, known_newly_created=False).dump(), volumes=res_volumes # type: ignore
|
|
||||||
)
|
|
||||||
|
|
||||||
for well, liquid_name, volume in zip(wells, liquid_names, volumes):
|
for well, liquid_name, volume in zip(wells, liquid_names, volumes):
|
||||||
well.set_liquids([(liquid_name, volume)]) # type: ignore
|
well.set_liquids([(liquid_name, volume)]) # type: ignore
|
||||||
|
res_samples.append({"name": well.name, "sample_uuid": well.unilabos_extra.get("sample_uuid", None)})
|
||||||
res_volumes.append(volume)
|
res_volumes.append(volume)
|
||||||
|
|
||||||
return SetLiquidReturn(
|
return SimpleReturn(samples=res_samples, volumes=res_volumes)
|
||||||
wells=ResourceTreeSet.from_plr_resources(wells, known_newly_created=False).dump(), volumes=res_volumes # type: ignore
|
|
||||||
)
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def set_liquid_from_plate(
|
|
||||||
cls, plate: ResourceSlot, well_names: list[str], liquid_names: list[str], volumes: list[float]
|
|
||||||
) -> SetLiquidFromPlateReturn:
|
|
||||||
"""Set the liquid in wells of a plate by well names (e.g., A1, A2, B3).
|
|
||||||
|
|
||||||
如果 liquid_names 和 volumes 为空,但 plate 和 well_names 不为空,直接返回 plate 和 wells。
|
|
||||||
"""
|
|
||||||
# 根据 well_names 获取对应的 Well 对象
|
|
||||||
wells = [plate.get_well(name) for name in well_names]
|
|
||||||
res_volumes = []
|
|
||||||
|
|
||||||
# 如果 liquid_names 和 volumes 都为空,直接返回
|
|
||||||
if not liquid_names and not volumes:
|
|
||||||
return SetLiquidFromPlateReturn(
|
|
||||||
plate=ResourceTreeSet.from_plr_resources([plate], known_newly_created=False).dump(), # type: ignore
|
|
||||||
wells=ResourceTreeSet.from_plr_resources(wells, known_newly_created=False).dump(), # type: ignore
|
|
||||||
volumes=res_volumes,
|
|
||||||
)
|
|
||||||
|
|
||||||
for well, liquid_name, volume in zip(wells, liquid_names, volumes):
|
|
||||||
well.set_liquids([(liquid_name, volume)]) # type: ignore
|
|
||||||
res_volumes.append(volume)
|
|
||||||
|
|
||||||
return SetLiquidFromPlateReturn(
|
|
||||||
plate=ResourceTreeSet.from_plr_resources([plate], known_newly_created=False).dump(), # type: ignore
|
|
||||||
wells=ResourceTreeSet.from_plr_resources(wells, known_newly_created=False).dump(), # type: ignore
|
|
||||||
volumes=res_volumes,
|
|
||||||
)
|
|
||||||
|
|
||||||
# ---------------------------------------------------------------
|
# ---------------------------------------------------------------
|
||||||
# REMOVE LIQUID --------------------------------------------------
|
# REMOVE LIQUID --------------------------------------------------
|
||||||
# ---------------------------------------------------------------
|
# ---------------------------------------------------------------
|
||||||
@@ -747,7 +676,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
target_rack = child
|
target_rack = child
|
||||||
target_rack = cast(TipRack, target_rack)
|
target_rack = cast(TipRack, target_rack)
|
||||||
available_tips = {}
|
available_tips = {}
|
||||||
for idx, tipSpot in enumerate(target_rack.get_all_items()):
|
for (idx, tipSpot) in enumerate(target_rack.get_all_items()):
|
||||||
if tipSpot.has_tip():
|
if tipSpot.has_tip():
|
||||||
available_tips[idx] = tipSpot
|
available_tips[idx] = tipSpot
|
||||||
continue
|
continue
|
||||||
@@ -755,8 +684,8 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
print("channel_num", self.channel_num)
|
print("channel_num", self.channel_num)
|
||||||
if self.channel_num == 8:
|
if self.channel_num == 8:
|
||||||
|
|
||||||
tip_prefix = list(available_tips.values())[0].name.split("_")[0]
|
tip_prefix = list(available_tips.values())[0].name.split('_')[0]
|
||||||
colnum_list = [int(tip.name.split("_")[-1][1:]) for tip in available_tips.values()]
|
colnum_list = [int(tip.name.split('_')[-1][1:]) for tip in available_tips.values()]
|
||||||
available_cols = [colnum for colnum, count in dict(Counter(colnum_list)).items() if count == 8]
|
available_cols = [colnum for colnum, count in dict(Counter(colnum_list)).items() if count == 8]
|
||||||
available_cols.sort()
|
available_cols.sort()
|
||||||
available_tips_dict = {tip.name: tip for tip in available_tips.values()}
|
available_tips_dict = {tip.name: tip for tip in available_tips.values()}
|
||||||
@@ -800,6 +729,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
"""Create a new protocol with the given metadata."""
|
"""Create a new protocol with the given metadata."""
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
async def remove_liquid(
|
async def remove_liquid(
|
||||||
self,
|
self,
|
||||||
vols: List[float],
|
vols: List[float],
|
||||||
@@ -858,11 +788,10 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
|
|
||||||
elif len(use_channels) == 8 and self.backend.num_channels == 8:
|
elif len(use_channels) == 8 and self.backend.num_channels == 8:
|
||||||
|
|
||||||
|
|
||||||
# 对于8个的情况,需要判断此时任务是不是能被8通道移液站来成功处理
|
# 对于8个的情况,需要判断此时任务是不是能被8通道移液站来成功处理
|
||||||
if len(sources) % 8 != 0:
|
if len(sources) % 8 != 0:
|
||||||
raise ValueError(
|
raise ValueError(f"Length of `sources` {len(sources)} must be a multiple of 8 for 8-channel mode.")
|
||||||
f"Length of `sources` {len(sources)} must be a multiple of 8 for 8-channel mode."
|
|
||||||
)
|
|
||||||
|
|
||||||
# 8个8个来取任务序列
|
# 8个8个来取任务序列
|
||||||
|
|
||||||
@@ -871,28 +800,18 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
for _ in range(len(use_channels)):
|
for _ in range(len(use_channels)):
|
||||||
tip.extend(next(self.current_tip))
|
tip.extend(next(self.current_tip))
|
||||||
await self.pick_up_tips(tip)
|
await self.pick_up_tips(tip)
|
||||||
current_targets = waste_liquid[i : i + 8]
|
current_targets = waste_liquid[i:i + 8]
|
||||||
current_reagent_sources = sources[i : i + 8]
|
current_reagent_sources = sources[i:i + 8]
|
||||||
current_asp_vols = vols[i : i + 8]
|
current_asp_vols = vols[i:i + 8]
|
||||||
current_dis_vols = vols[i : i + 8]
|
current_dis_vols = vols[i:i + 8]
|
||||||
current_asp_flow_rates = flow_rates[i : i + 8] if flow_rates else [None] * 8
|
current_asp_flow_rates = flow_rates[i:i + 8] if flow_rates else [None] * 8
|
||||||
current_dis_flow_rates = (
|
current_dis_flow_rates = flow_rates[-i*8-8:len(flow_rates)-i*8] if flow_rates else [None] * 8
|
||||||
flow_rates[-i * 8 - 8 : len(flow_rates) - i * 8] if flow_rates else [None] * 8
|
current_asp_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
)
|
current_dis_offset = offsets[-i*8-8:len(offsets)-i*8] if offsets else [None] * 8
|
||||||
current_asp_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_asp_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_dis_offset = offsets[-i * 8 - 8 : len(offsets) - i * 8] if offsets else [None] * 8
|
current_dis_liquid_height = liquid_height[-i*8-8:len(liquid_height)-i*8] if liquid_height else [None] * 8
|
||||||
current_asp_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_asp_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
current_dis_liquid_height = (
|
current_dis_blow_out_air_volume = blow_out_air_volume[-i*8-8:len(blow_out_air_volume)-i*8] if blow_out_air_volume else [None] * 8
|
||||||
liquid_height[-i * 8 - 8 : len(liquid_height) - i * 8] if liquid_height else [None] * 8
|
|
||||||
)
|
|
||||||
current_asp_blow_out_air_volume = (
|
|
||||||
blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
|
||||||
)
|
|
||||||
current_dis_blow_out_air_volume = (
|
|
||||||
blow_out_air_volume[-i * 8 - 8 : len(blow_out_air_volume) - i * 8]
|
|
||||||
if blow_out_air_volume
|
|
||||||
else [None] * 8
|
|
||||||
)
|
|
||||||
|
|
||||||
await self.aspirate(
|
await self.aspirate(
|
||||||
resources=current_reagent_sources,
|
resources=current_reagent_sources,
|
||||||
@@ -1017,28 +936,18 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
for _ in range(len(use_channels)):
|
for _ in range(len(use_channels)):
|
||||||
tip.extend(next(self.current_tip))
|
tip.extend(next(self.current_tip))
|
||||||
await self.pick_up_tips(tip)
|
await self.pick_up_tips(tip)
|
||||||
current_targets = targets[i : i + 8]
|
current_targets = targets[i:i + 8]
|
||||||
current_reagent_sources = reagent_sources[i : i + 8]
|
current_reagent_sources = reagent_sources[i:i + 8]
|
||||||
current_asp_vols = asp_vols[i : i + 8]
|
current_asp_vols = asp_vols[i:i + 8]
|
||||||
current_dis_vols = dis_vols[i : i + 8]
|
current_dis_vols = dis_vols[i:i + 8]
|
||||||
current_asp_flow_rates = flow_rates[i : i + 8] if flow_rates else [None] * 8
|
current_asp_flow_rates = flow_rates[i:i + 8] if flow_rates else [None] * 8
|
||||||
current_dis_flow_rates = (
|
current_dis_flow_rates = flow_rates[-i*8-8:len(flow_rates)-i*8] if flow_rates else [None] * 8
|
||||||
flow_rates[-i * 8 - 8 : len(flow_rates) - i * 8] if flow_rates else [None] * 8
|
current_asp_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
)
|
current_dis_offset = offsets[-i*8-8:len(offsets)-i*8] if offsets else [None] * 8
|
||||||
current_asp_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_asp_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_dis_offset = offsets[-i * 8 - 8 : len(offsets) - i * 8] if offsets else [None] * 8
|
current_dis_liquid_height = liquid_height[-i*8-8:len(liquid_height)-i*8] if liquid_height else [None] * 8
|
||||||
current_asp_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_asp_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
current_dis_liquid_height = (
|
current_dis_blow_out_air_volume = blow_out_air_volume[-i*8-8:len(blow_out_air_volume)-i*8] if blow_out_air_volume else [None] * 8
|
||||||
liquid_height[-i * 8 - 8 : len(liquid_height) - i * 8] if liquid_height else [None] * 8
|
|
||||||
)
|
|
||||||
current_asp_blow_out_air_volume = (
|
|
||||||
blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
|
||||||
)
|
|
||||||
current_dis_blow_out_air_volume = (
|
|
||||||
blow_out_air_volume[-i * 8 - 8 : len(blow_out_air_volume) - i * 8]
|
|
||||||
if blow_out_air_volume
|
|
||||||
else [None] * 8
|
|
||||||
)
|
|
||||||
|
|
||||||
await self.aspirate(
|
await self.aspirate(
|
||||||
resources=current_reagent_sources,
|
resources=current_reagent_sources,
|
||||||
@@ -1080,6 +989,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
await self.touch_tip(current_targets)
|
await self.touch_tip(current_targets)
|
||||||
await self.discard_tips()
|
await self.discard_tips()
|
||||||
|
|
||||||
|
|
||||||
# except Exception as e:
|
# except Exception as e:
|
||||||
# traceback.print_exc()
|
# traceback.print_exc()
|
||||||
# raise RuntimeError(f"Liquid addition failed: {e}") from e
|
# raise RuntimeError(f"Liquid addition failed: {e}") from e
|
||||||
@@ -1179,71 +1089,29 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
if num_sources == 1 and num_targets > 1:
|
if num_sources == 1 and num_targets > 1:
|
||||||
# 模式1: 一对多 (1 source -> N targets)
|
# 模式1: 一对多 (1 source -> N targets)
|
||||||
await self._transfer_one_to_many(
|
await self._transfer_one_to_many(
|
||||||
sources[0],
|
sources[0], targets, tip_racks, use_channels,
|
||||||
targets,
|
asp_vols, dis_vols, asp_flow_rates, dis_flow_rates,
|
||||||
tip_racks,
|
offsets, touch_tip, liquid_height, blow_out_air_volume,
|
||||||
use_channels,
|
spread, mix_stage, mix_times, mix_vol, mix_rate,
|
||||||
asp_vols,
|
mix_liquid_height, delays
|
||||||
dis_vols,
|
|
||||||
asp_flow_rates,
|
|
||||||
dis_flow_rates,
|
|
||||||
offsets,
|
|
||||||
touch_tip,
|
|
||||||
liquid_height,
|
|
||||||
blow_out_air_volume,
|
|
||||||
spread,
|
|
||||||
mix_stage,
|
|
||||||
mix_times,
|
|
||||||
mix_vol,
|
|
||||||
mix_rate,
|
|
||||||
mix_liquid_height,
|
|
||||||
delays,
|
|
||||||
)
|
)
|
||||||
elif num_sources > 1 and num_targets == 1:
|
elif num_sources > 1 and num_targets == 1:
|
||||||
# 模式2: 多对一 (N sources -> 1 target)
|
# 模式2: 多对一 (N sources -> 1 target)
|
||||||
await self._transfer_many_to_one(
|
await self._transfer_many_to_one(
|
||||||
sources,
|
sources, targets[0], tip_racks, use_channels,
|
||||||
targets[0],
|
asp_vols, dis_vols, asp_flow_rates, dis_flow_rates,
|
||||||
tip_racks,
|
offsets, touch_tip, liquid_height, blow_out_air_volume,
|
||||||
use_channels,
|
spread, mix_stage, mix_times, mix_vol, mix_rate,
|
||||||
asp_vols,
|
mix_liquid_height, delays
|
||||||
dis_vols,
|
|
||||||
asp_flow_rates,
|
|
||||||
dis_flow_rates,
|
|
||||||
offsets,
|
|
||||||
touch_tip,
|
|
||||||
liquid_height,
|
|
||||||
blow_out_air_volume,
|
|
||||||
spread,
|
|
||||||
mix_stage,
|
|
||||||
mix_times,
|
|
||||||
mix_vol,
|
|
||||||
mix_rate,
|
|
||||||
mix_liquid_height,
|
|
||||||
delays,
|
|
||||||
)
|
)
|
||||||
elif num_sources == num_targets:
|
elif num_sources == num_targets:
|
||||||
# 模式3: 一对一 (N sources -> N targets)
|
# 模式3: 一对一 (N sources -> N targets)
|
||||||
await self._transfer_one_to_one(
|
await self._transfer_one_to_one(
|
||||||
sources,
|
sources, targets, tip_racks, use_channels,
|
||||||
targets,
|
asp_vols, dis_vols, asp_flow_rates, dis_flow_rates,
|
||||||
tip_racks,
|
offsets, touch_tip, liquid_height, blow_out_air_volume,
|
||||||
use_channels,
|
spread, mix_stage, mix_times, mix_vol, mix_rate,
|
||||||
asp_vols,
|
mix_liquid_height, delays
|
||||||
dis_vols,
|
|
||||||
asp_flow_rates,
|
|
||||||
dis_flow_rates,
|
|
||||||
offsets,
|
|
||||||
touch_tip,
|
|
||||||
liquid_height,
|
|
||||||
blow_out_air_volume,
|
|
||||||
spread,
|
|
||||||
mix_stage,
|
|
||||||
mix_times,
|
|
||||||
mix_vol,
|
|
||||||
mix_rate,
|
|
||||||
mix_liquid_height,
|
|
||||||
delays,
|
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
@@ -1306,9 +1174,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
flow_rates=[asp_flow_rates[_]] if asp_flow_rates and len(asp_flow_rates) > _ else None,
|
flow_rates=[asp_flow_rates[_]] if asp_flow_rates and len(asp_flow_rates) > _ else None,
|
||||||
offsets=[offsets[_]] if offsets and len(offsets) > _ else None,
|
offsets=[offsets[_]] if offsets and len(offsets) > _ else None,
|
||||||
liquid_height=[liquid_height[_]] if liquid_height and len(liquid_height) > _ else None,
|
liquid_height=[liquid_height[_]] if liquid_height and len(liquid_height) > _ else None,
|
||||||
blow_out_air_volume=(
|
blow_out_air_volume=[blow_out_air_volume[_]] if blow_out_air_volume and len(blow_out_air_volume) > _ else None,
|
||||||
[blow_out_air_volume[_]] if blow_out_air_volume and len(blow_out_air_volume) > _ else None
|
|
||||||
),
|
|
||||||
spread=spread,
|
spread=spread,
|
||||||
)
|
)
|
||||||
if delays is not None:
|
if delays is not None:
|
||||||
@@ -1319,9 +1185,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
use_channels=use_channels,
|
use_channels=use_channels,
|
||||||
flow_rates=[dis_flow_rates[_]] if dis_flow_rates and len(dis_flow_rates) > _ else None,
|
flow_rates=[dis_flow_rates[_]] if dis_flow_rates and len(dis_flow_rates) > _ else None,
|
||||||
offsets=[offsets[_]] if offsets and len(offsets) > _ else None,
|
offsets=[offsets[_]] if offsets and len(offsets) > _ else None,
|
||||||
blow_out_air_volume=(
|
blow_out_air_volume=[blow_out_air_volume[_]] if blow_out_air_volume and len(blow_out_air_volume) > _ else None,
|
||||||
[blow_out_air_volume[_]] if blow_out_air_volume and len(blow_out_air_volume) > _ else None
|
|
||||||
),
|
|
||||||
liquid_height=[liquid_height[_]] if liquid_height and len(liquid_height) > _ else None,
|
liquid_height=[liquid_height[_]] if liquid_height and len(liquid_height) > _ else None,
|
||||||
spread=spread,
|
spread=spread,
|
||||||
)
|
)
|
||||||
@@ -1350,18 +1214,18 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
for _ in range(len(use_channels)):
|
for _ in range(len(use_channels)):
|
||||||
tip.extend(next(self.current_tip))
|
tip.extend(next(self.current_tip))
|
||||||
await self.pick_up_tips(tip)
|
await self.pick_up_tips(tip)
|
||||||
current_targets = targets[i : i + 8]
|
current_targets = targets[i:i + 8]
|
||||||
current_reagent_sources = sources[i : i + 8]
|
current_reagent_sources = sources[i:i + 8]
|
||||||
current_asp_vols = asp_vols[i : i + 8]
|
current_asp_vols = asp_vols[i:i + 8]
|
||||||
current_dis_vols = dis_vols[i : i + 8]
|
current_dis_vols = dis_vols[i:i + 8]
|
||||||
current_asp_flow_rates = asp_flow_rates[i : i + 8] if asp_flow_rates else None
|
current_asp_flow_rates = asp_flow_rates[i:i + 8] if asp_flow_rates else None
|
||||||
current_asp_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_asp_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
current_dis_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_dis_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
current_asp_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_asp_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_dis_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_dis_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_asp_blow_out_air_volume = blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
current_asp_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
current_dis_blow_out_air_volume = blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
current_dis_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
current_dis_flow_rates = dis_flow_rates[i : i + 8] if dis_flow_rates else None
|
current_dis_flow_rates = dis_flow_rates[i:i + 8] if dis_flow_rates else None
|
||||||
|
|
||||||
if mix_stage in ["before", "both"] and mix_times is not None and mix_times > 0:
|
if mix_stage in ["before", "both"] and mix_times is not None and mix_times > 0:
|
||||||
await self.mix(
|
await self.mix(
|
||||||
@@ -1411,7 +1275,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
if delays is not None and len(delays) > 1:
|
if delays is not None and len(delays) > 1:
|
||||||
await self.custom_delay(seconds=delays[1])
|
await self.custom_delay(seconds=delays[1])
|
||||||
await self.touch_tip(current_targets)
|
await self.touch_tip(current_targets)
|
||||||
await self.discard_tips([0, 1, 2, 3, 4, 5, 6, 7])
|
await self.discard_tips([0,1,2,3,4,5,6,7])
|
||||||
|
|
||||||
async def _transfer_one_to_many(
|
async def _transfer_one_to_many(
|
||||||
self,
|
self,
|
||||||
@@ -1460,7 +1324,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
targets=[target],
|
targets=[target],
|
||||||
mix_time=mix_times,
|
mix_time=mix_times,
|
||||||
mix_vol=mix_vol,
|
mix_vol=mix_vol,
|
||||||
offsets=offsets[idx : idx + 1] if offsets and len(offsets) > idx else None,
|
offsets=offsets[idx:idx + 1] if offsets and len(offsets) > idx else None,
|
||||||
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
||||||
mix_rate=mix_rate if mix_rate else None,
|
mix_rate=mix_rate if mix_rate else None,
|
||||||
)
|
)
|
||||||
@@ -1473,9 +1337,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
flow_rates=[asp_flow_rates[0]] if asp_flow_rates and len(asp_flow_rates) > 0 else None,
|
flow_rates=[asp_flow_rates[0]] if asp_flow_rates and len(asp_flow_rates) > 0 else None,
|
||||||
offsets=[offsets[0]] if offsets and len(offsets) > 0 else None,
|
offsets=[offsets[0]] if offsets and len(offsets) > 0 else None,
|
||||||
liquid_height=[liquid_height[0]] if liquid_height and len(liquid_height) > 0 else None,
|
liquid_height=[liquid_height[0]] if liquid_height and len(liquid_height) > 0 else None,
|
||||||
blow_out_air_volume=(
|
blow_out_air_volume=[blow_out_air_volume[0]] if blow_out_air_volume and len(blow_out_air_volume) > 0 else None,
|
||||||
[blow_out_air_volume[0]] if blow_out_air_volume and len(blow_out_air_volume) > 0 else None
|
|
||||||
),
|
|
||||||
spread=spread,
|
spread=spread,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -1490,9 +1352,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
use_channels=use_channels,
|
use_channels=use_channels,
|
||||||
flow_rates=[dis_flow_rates[idx]] if dis_flow_rates and len(dis_flow_rates) > idx else None,
|
flow_rates=[dis_flow_rates[idx]] if dis_flow_rates and len(dis_flow_rates) > idx else None,
|
||||||
offsets=[offsets[idx]] if offsets and len(offsets) > idx else None,
|
offsets=[offsets[idx]] if offsets and len(offsets) > idx else None,
|
||||||
blow_out_air_volume=(
|
blow_out_air_volume=[blow_out_air_volume[idx]] if blow_out_air_volume and len(blow_out_air_volume) > idx else None,
|
||||||
[blow_out_air_volume[idx]] if blow_out_air_volume and len(blow_out_air_volume) > idx else None
|
|
||||||
),
|
|
||||||
liquid_height=[liquid_height[idx]] if liquid_height and len(liquid_height) > idx else None,
|
liquid_height=[liquid_height[idx]] if liquid_height and len(liquid_height) > idx else None,
|
||||||
spread=spread,
|
spread=spread,
|
||||||
)
|
)
|
||||||
@@ -1503,7 +1363,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
targets=[target],
|
targets=[target],
|
||||||
mix_time=mix_times,
|
mix_time=mix_times,
|
||||||
mix_vol=mix_vol,
|
mix_vol=mix_vol,
|
||||||
offsets=offsets[idx : idx + 1] if offsets else None,
|
offsets=offsets[idx:idx+1] if offsets else None,
|
||||||
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
||||||
mix_rate=mix_rate if mix_rate else None,
|
mix_rate=mix_rate if mix_rate else None,
|
||||||
)
|
)
|
||||||
@@ -1524,29 +1384,21 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
tip.extend(next(self.current_tip))
|
tip.extend(next(self.current_tip))
|
||||||
await self.pick_up_tips(tip)
|
await self.pick_up_tips(tip)
|
||||||
|
|
||||||
current_targets = targets[i : i + 8]
|
current_targets = targets[i:i + 8]
|
||||||
current_dis_vols = dis_vols[i : i + 8]
|
current_dis_vols = dis_vols[i:i + 8]
|
||||||
|
|
||||||
# 8个通道都从同一个源容器吸液,每个通道的吸液体积等于对应的分液体积
|
# 8个通道都从同一个源容器吸液,每个通道的吸液体积等于对应的分液体积
|
||||||
current_asp_flow_rates = (
|
current_asp_flow_rates = asp_flow_rates[0:1] * 8 if asp_flow_rates and len(asp_flow_rates) > 0 else None
|
||||||
asp_flow_rates[0:1] * 8 if asp_flow_rates and len(asp_flow_rates) > 0 else None
|
|
||||||
)
|
|
||||||
current_asp_offset = offsets[0:1] * 8 if offsets and len(offsets) > 0 else [None] * 8
|
current_asp_offset = offsets[0:1] * 8 if offsets and len(offsets) > 0 else [None] * 8
|
||||||
current_asp_liquid_height = (
|
current_asp_liquid_height = liquid_height[0:1] * 8 if liquid_height and len(liquid_height) > 0 else [None] * 8
|
||||||
liquid_height[0:1] * 8 if liquid_height and len(liquid_height) > 0 else [None] * 8
|
current_asp_blow_out_air_volume = blow_out_air_volume[0:1] * 8 if blow_out_air_volume and len(blow_out_air_volume) > 0 else [None] * 8
|
||||||
)
|
|
||||||
current_asp_blow_out_air_volume = (
|
|
||||||
blow_out_air_volume[0:1] * 8
|
|
||||||
if blow_out_air_volume and len(blow_out_air_volume) > 0
|
|
||||||
else [None] * 8
|
|
||||||
)
|
|
||||||
|
|
||||||
if mix_stage in ["before", "both"] and mix_times is not None and mix_times > 0:
|
if mix_stage in ["before", "both"] and mix_times is not None and mix_times > 0:
|
||||||
await self.mix(
|
await self.mix(
|
||||||
targets=current_targets,
|
targets=current_targets,
|
||||||
mix_time=mix_times,
|
mix_time=mix_times,
|
||||||
mix_vol=mix_vol,
|
mix_vol=mix_vol,
|
||||||
offsets=offsets[i : i + 8] if offsets else None,
|
offsets=offsets[i:i + 8] if offsets else None,
|
||||||
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
height_to_bottom=mix_liquid_height if mix_liquid_height else None,
|
||||||
mix_rate=mix_rate if mix_rate else None,
|
mix_rate=mix_rate if mix_rate else None,
|
||||||
)
|
)
|
||||||
@@ -1567,10 +1419,10 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
await self.custom_delay(seconds=delays[0])
|
await self.custom_delay(seconds=delays[0])
|
||||||
|
|
||||||
# 分液到8个目标
|
# 分液到8个目标
|
||||||
current_dis_flow_rates = dis_flow_rates[i : i + 8] if dis_flow_rates else None
|
current_dis_flow_rates = dis_flow_rates[i:i + 8] if dis_flow_rates else None
|
||||||
current_dis_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_dis_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
current_dis_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_dis_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_dis_blow_out_air_volume = blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
current_dis_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
|
|
||||||
await self.dispense(
|
await self.dispense(
|
||||||
resources=current_targets,
|
resources=current_targets,
|
||||||
@@ -1599,7 +1451,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
if touch_tip:
|
if touch_tip:
|
||||||
await self.touch_tip(current_targets)
|
await self.touch_tip(current_targets)
|
||||||
|
|
||||||
await self.discard_tips([0, 1, 2, 3, 4, 5, 6, 7])
|
await self.discard_tips([0,1,2,3,4,5,6,7])
|
||||||
|
|
||||||
async def _transfer_many_to_one(
|
async def _transfer_many_to_one(
|
||||||
self,
|
self,
|
||||||
@@ -1672,9 +1524,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
flow_rates=[asp_flow_rates[idx]] if asp_flow_rates and len(asp_flow_rates) > idx else None,
|
flow_rates=[asp_flow_rates[idx]] if asp_flow_rates and len(asp_flow_rates) > idx else None,
|
||||||
offsets=[offsets[idx]] if offsets and len(offsets) > idx else None,
|
offsets=[offsets[idx]] if offsets and len(offsets) > idx else None,
|
||||||
liquid_height=[liquid_height[idx]] if liquid_height and len(liquid_height) > idx else None,
|
liquid_height=[liquid_height[idx]] if liquid_height and len(liquid_height) > idx else None,
|
||||||
blow_out_air_volume=(
|
blow_out_air_volume=[blow_out_air_volume[idx]] if blow_out_air_volume and len(blow_out_air_volume) > idx else None,
|
||||||
[blow_out_air_volume[idx]] if blow_out_air_volume and len(blow_out_air_volume) > idx else None
|
|
||||||
),
|
|
||||||
spread=spread,
|
spread=spread,
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -1688,18 +1538,14 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
dis_flow_rate = dis_flow_rates[idx] if dis_flow_rates and len(dis_flow_rates) > idx else None
|
dis_flow_rate = dis_flow_rates[idx] if dis_flow_rates and len(dis_flow_rates) > idx else None
|
||||||
dis_offset = offsets[idx] if offsets and len(offsets) > idx else None
|
dis_offset = offsets[idx] if offsets and len(offsets) > idx else None
|
||||||
dis_liquid_height = liquid_height[idx] if liquid_height and len(liquid_height) > idx else None
|
dis_liquid_height = liquid_height[idx] if liquid_height and len(liquid_height) > idx else None
|
||||||
dis_blow_out = (
|
dis_blow_out = blow_out_air_volume[idx] if blow_out_air_volume and len(blow_out_air_volume) > idx else None
|
||||||
blow_out_air_volume[idx] if blow_out_air_volume and len(blow_out_air_volume) > idx else None
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
# 标准模式:分液体积等于吸液体积
|
# 标准模式:分液体积等于吸液体积
|
||||||
dis_vol = asp_vols[idx]
|
dis_vol = asp_vols[idx]
|
||||||
dis_flow_rate = dis_flow_rates[0] if dis_flow_rates and len(dis_flow_rates) > 0 else None
|
dis_flow_rate = dis_flow_rates[0] if dis_flow_rates and len(dis_flow_rates) > 0 else None
|
||||||
dis_offset = offsets[0] if offsets and len(offsets) > 0 else None
|
dis_offset = offsets[0] if offsets and len(offsets) > 0 else None
|
||||||
dis_liquid_height = liquid_height[0] if liquid_height and len(liquid_height) > 0 else None
|
dis_liquid_height = liquid_height[0] if liquid_height and len(liquid_height) > 0 else None
|
||||||
dis_blow_out = (
|
dis_blow_out = blow_out_air_volume[0] if blow_out_air_volume and len(blow_out_air_volume) > 0 else None
|
||||||
blow_out_air_volume[0] if blow_out_air_volume and len(blow_out_air_volume) > 0 else None
|
|
||||||
)
|
|
||||||
|
|
||||||
await self.dispense(
|
await self.dispense(
|
||||||
resources=[target],
|
resources=[target],
|
||||||
@@ -1753,12 +1599,12 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
tip.extend(next(self.current_tip))
|
tip.extend(next(self.current_tip))
|
||||||
await self.pick_up_tips(tip)
|
await self.pick_up_tips(tip)
|
||||||
|
|
||||||
current_sources = sources[i : i + 8]
|
current_sources = sources[i:i + 8]
|
||||||
current_asp_vols = asp_vols[i : i + 8]
|
current_asp_vols = asp_vols[i:i + 8]
|
||||||
current_asp_flow_rates = asp_flow_rates[i : i + 8] if asp_flow_rates else None
|
current_asp_flow_rates = asp_flow_rates[i:i + 8] if asp_flow_rates else None
|
||||||
current_asp_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_asp_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
current_asp_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_asp_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_asp_blow_out_air_volume = blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
current_asp_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
|
|
||||||
# 从8个源容器吸液
|
# 从8个源容器吸液
|
||||||
await self.aspirate(
|
await self.aspirate(
|
||||||
@@ -1778,22 +1624,18 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
# 分液到目标容器(每个通道分液到同一个目标)
|
# 分液到目标容器(每个通道分液到同一个目标)
|
||||||
if use_proportional_mixing:
|
if use_proportional_mixing:
|
||||||
# 按比例混合:使用对应的 dis_vols
|
# 按比例混合:使用对应的 dis_vols
|
||||||
current_dis_vols = dis_vols[i : i + 8]
|
current_dis_vols = dis_vols[i:i + 8]
|
||||||
current_dis_flow_rates = dis_flow_rates[i : i + 8] if dis_flow_rates else None
|
current_dis_flow_rates = dis_flow_rates[i:i + 8] if dis_flow_rates else None
|
||||||
current_dis_offset = offsets[i : i + 8] if offsets else [None] * 8
|
current_dis_offset = offsets[i:i + 8] if offsets else [None] * 8
|
||||||
current_dis_liquid_height = liquid_height[i : i + 8] if liquid_height else [None] * 8
|
current_dis_liquid_height = liquid_height[i:i + 8] if liquid_height else [None] * 8
|
||||||
current_dis_blow_out_air_volume = (
|
current_dis_blow_out_air_volume = blow_out_air_volume[i:i + 8] if blow_out_air_volume else [None] * 8
|
||||||
blow_out_air_volume[i : i + 8] if blow_out_air_volume else [None] * 8
|
|
||||||
)
|
|
||||||
else:
|
else:
|
||||||
# 标准模式:每个通道分液体积等于其吸液体积
|
# 标准模式:每个通道分液体积等于其吸液体积
|
||||||
current_dis_vols = current_asp_vols
|
current_dis_vols = current_asp_vols
|
||||||
current_dis_flow_rates = dis_flow_rates[0:1] * 8 if dis_flow_rates else None
|
current_dis_flow_rates = dis_flow_rates[0:1] * 8 if dis_flow_rates else None
|
||||||
current_dis_offset = offsets[0:1] * 8 if offsets else [None] * 8
|
current_dis_offset = offsets[0:1] * 8 if offsets else [None] * 8
|
||||||
current_dis_liquid_height = liquid_height[0:1] * 8 if liquid_height else [None] * 8
|
current_dis_liquid_height = liquid_height[0:1] * 8 if liquid_height else [None] * 8
|
||||||
current_dis_blow_out_air_volume = (
|
current_dis_blow_out_air_volume = blow_out_air_volume[0:1] * 8 if blow_out_air_volume else [None] * 8
|
||||||
blow_out_air_volume[0:1] * 8 if blow_out_air_volume else [None] * 8
|
|
||||||
)
|
|
||||||
|
|
||||||
await self.dispense(
|
await self.dispense(
|
||||||
resources=[target] * 8, # 8个通道都分到同一个目标
|
resources=[target] * 8, # 8个通道都分到同一个目标
|
||||||
@@ -1809,7 +1651,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
if delays is not None and len(delays) > 1:
|
if delays is not None and len(delays) > 1:
|
||||||
await self.custom_delay(seconds=delays[1])
|
await self.custom_delay(seconds=delays[1])
|
||||||
|
|
||||||
await self.discard_tips([0, 1, 2, 3, 4, 5, 6, 7])
|
await self.discard_tips([0,1,2,3,4,5,6,7])
|
||||||
|
|
||||||
# 最后在目标容器中混合(如果需要)
|
# 最后在目标容器中混合(如果需要)
|
||||||
if mix_stage in ["after", "both"] and mix_times is not None and mix_times > 0:
|
if mix_stage in ["after", "both"] and mix_times is not None and mix_times > 0:
|
||||||
@@ -1829,6 +1671,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
# traceback.print_exc()
|
# traceback.print_exc()
|
||||||
# raise RuntimeError(f"Liquid addition failed: {e}") from e
|
# raise RuntimeError(f"Liquid addition failed: {e}") from e
|
||||||
|
|
||||||
|
|
||||||
# ---------------------------------------------------------------
|
# ---------------------------------------------------------------
|
||||||
# Helper utilities
|
# Helper utilities
|
||||||
# ---------------------------------------------------------------
|
# ---------------------------------------------------------------
|
||||||
@@ -1849,6 +1692,7 @@ class LiquidHandlerAbstract(LiquidHandlerMiddleware):
|
|||||||
print(f"Current time: {time.strftime('%H:%M:%S')}")
|
print(f"Current time: {time.strftime('%H:%M:%S')}")
|
||||||
|
|
||||||
async def touch_tip(self, targets: Sequence[Container]):
|
async def touch_tip(self, targets: Sequence[Container]):
|
||||||
|
|
||||||
"""Touch the tip to the side of the well."""
|
"""Touch the tip to the side of the well."""
|
||||||
|
|
||||||
if not self.support_touch_tip:
|
if not self.support_touch_tip:
|
||||||
|
|||||||
@@ -30,30 +30,9 @@ from pylabrobot.liquid_handling.standard import (
|
|||||||
ResourceMove,
|
ResourceMove,
|
||||||
ResourceDrop,
|
ResourceDrop,
|
||||||
)
|
)
|
||||||
from pylabrobot.resources import (
|
from pylabrobot.resources import ResourceHolder, ResourceStack, Tip, Deck, Plate, Well, TipRack, Resource, Container, Coordinate, TipSpot, Trash, PlateAdapter, TubeRack
|
||||||
ResourceHolder,
|
|
||||||
ResourceStack,
|
|
||||||
Tip,
|
|
||||||
Deck,
|
|
||||||
Plate,
|
|
||||||
Well,
|
|
||||||
TipRack,
|
|
||||||
Resource,
|
|
||||||
Container,
|
|
||||||
Coordinate,
|
|
||||||
TipSpot,
|
|
||||||
Trash,
|
|
||||||
PlateAdapter,
|
|
||||||
TubeRack,
|
|
||||||
)
|
|
||||||
|
|
||||||
from unilabos.devices.liquid_handling.liquid_handler_abstract import (
|
from unilabos.devices.liquid_handling.liquid_handler_abstract import LiquidHandlerAbstract, SimpleReturn
|
||||||
LiquidHandlerAbstract,
|
|
||||||
SimpleReturn,
|
|
||||||
SetLiquidReturn,
|
|
||||||
SetLiquidFromPlateReturn,
|
|
||||||
)
|
|
||||||
from unilabos.registry.placeholder_type import ResourceSlot
|
|
||||||
from unilabos.ros.nodes.base_device_node import BaseROS2DeviceNode
|
from unilabos.ros.nodes.base_device_node import BaseROS2DeviceNode
|
||||||
|
|
||||||
|
|
||||||
@@ -101,7 +80,6 @@ class PRCXI9300Deck(Deck):
|
|||||||
self.slots[slot - 1] = resource
|
self.slots[slot - 1] = resource
|
||||||
super().assign_child_resource(resource, location=self.slot_locations[slot - 1])
|
super().assign_child_resource(resource, location=self.slot_locations[slot - 1])
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300Container(Plate):
|
class PRCXI9300Container(Plate):
|
||||||
"""PRCXI 9300 的专用 Container 类,继承自 Plate,用于槽位定位和未知模块。
|
"""PRCXI 9300 的专用 Container 类,继承自 Plate,用于槽位定位和未知模块。
|
||||||
|
|
||||||
@@ -131,28 +109,19 @@ class PRCXI9300Container(Plate):
|
|||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
data.update(self._unilabos_state)
|
data.update(self._unilabos_state)
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300Plate(Plate):
|
class PRCXI9300Plate(Plate):
|
||||||
"""
|
"""
|
||||||
专用孔板类:
|
专用孔板类:
|
||||||
1. 继承自 PLR 原生 Plate,保留所有物理特性。
|
1. 继承自 PLR 原生 Plate,保留所有物理特性。
|
||||||
2. 增加 material_info 参数,用于在初始化时直接绑定 Unilab UUID。
|
2. 增加 material_info 参数,用于在初始化时直接绑定 Unilab UUID。
|
||||||
"""
|
"""
|
||||||
|
def __init__(self, name: str, size_x: float, size_y: float, size_z: float,
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
name: str,
|
|
||||||
size_x: float,
|
|
||||||
size_y: float,
|
|
||||||
size_z: float,
|
|
||||||
category: str = "plate",
|
category: str = "plate",
|
||||||
ordered_items: collections.OrderedDict = None,
|
ordered_items: collections.OrderedDict = None,
|
||||||
ordering: Optional[collections.OrderedDict] = None,
|
ordering: Optional[collections.OrderedDict] = None,
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
material_info: Optional[Dict[str, Any]] = None,
|
material_info: Optional[Dict[str, Any]] = None,
|
||||||
**kwargs,
|
**kwargs):
|
||||||
):
|
|
||||||
# 如果 ordered_items 不为 None,直接使用
|
# 如果 ordered_items 不为 None,直接使用
|
||||||
if ordered_items is not None:
|
if ordered_items is not None:
|
||||||
items = ordered_items
|
items = ordered_items
|
||||||
@@ -176,31 +145,37 @@ class PRCXI9300Plate(Plate):
|
|||||||
|
|
||||||
# 根据情况传递不同的参数
|
# 根据情况传递不同的参数
|
||||||
if items is not None:
|
if items is not None:
|
||||||
super().__init__(
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
name, size_x, size_y, size_z, ordered_items=items, category=category, model=model, **kwargs
|
ordered_items=items,
|
||||||
)
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
elif ordering_param is not None:
|
elif ordering_param is not None:
|
||||||
# 传递 ordering 参数,让 Plate 自己创建 Well 对象
|
# 传递 ordering 参数,让 Plate 自己创建 Well 对象
|
||||||
super().__init__(
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
name, size_x, size_y, size_z, ordering=ordering_param, category=category, model=model, **kwargs
|
ordering=ordering_param,
|
||||||
)
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
else:
|
else:
|
||||||
super().__init__(name, size_x, size_y, size_z, category=category, model=model, **kwargs)
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
|
|
||||||
self._unilabos_state = {}
|
self._unilabos_state = {}
|
||||||
if material_info:
|
if material_info:
|
||||||
self._unilabos_state["Material"] = material_info
|
self._unilabos_state["Material"] = material_info
|
||||||
|
|
||||||
|
|
||||||
def load_state(self, state: Dict[str, Any]) -> None:
|
def load_state(self, state: Dict[str, Any]) -> None:
|
||||||
super().load_state(state)
|
super().load_state(state)
|
||||||
self._unilabos_state = state
|
self._unilabos_state = state
|
||||||
|
|
||||||
|
|
||||||
def serialize_state(self) -> Dict[str, Dict[str, Any]]:
|
def serialize_state(self) -> Dict[str, Dict[str, Any]]:
|
||||||
try:
|
try:
|
||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
data = {}
|
data = {}
|
||||||
if hasattr(self, "_unilabos_state") and self._unilabos_state:
|
if hasattr(self, '_unilabos_state') and self._unilabos_state:
|
||||||
safe_state = {}
|
safe_state = {}
|
||||||
for k, v in self._unilabos_state.items():
|
for k, v in self._unilabos_state.items():
|
||||||
# 如果是 Material 字典,深入检查
|
# 如果是 Material 字典,深入检查
|
||||||
@@ -221,24 +196,15 @@ class PRCXI9300Plate(Plate):
|
|||||||
|
|
||||||
data.update(safe_state)
|
data.update(safe_state)
|
||||||
return data # 其他顶层属性也进行类型检查
|
return data # 其他顶层属性也进行类型检查
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300TipRack(TipRack):
|
class PRCXI9300TipRack(TipRack):
|
||||||
"""专用吸头盒类"""
|
""" 专用吸头盒类 """
|
||||||
|
def __init__(self, name: str, size_x: float, size_y: float, size_z: float,
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
name: str,
|
|
||||||
size_x: float,
|
|
||||||
size_y: float,
|
|
||||||
size_z: float,
|
|
||||||
category: str = "tip_rack",
|
category: str = "tip_rack",
|
||||||
ordered_items: collections.OrderedDict = None,
|
ordered_items: collections.OrderedDict = None,
|
||||||
ordering: Optional[collections.OrderedDict] = None,
|
ordering: Optional[collections.OrderedDict] = None,
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
material_info: Optional[Dict[str, Any]] = None,
|
material_info: Optional[Dict[str, Any]] = None,
|
||||||
**kwargs,
|
**kwargs):
|
||||||
):
|
|
||||||
# 如果 ordered_items 不为 None,直接使用
|
# 如果 ordered_items 不为 None,直接使用
|
||||||
if ordered_items is not None:
|
if ordered_items is not None:
|
||||||
items = ordered_items
|
items = ordered_items
|
||||||
@@ -262,16 +228,20 @@ class PRCXI9300TipRack(TipRack):
|
|||||||
|
|
||||||
# 根据情况传递不同的参数
|
# 根据情况传递不同的参数
|
||||||
if items is not None:
|
if items is not None:
|
||||||
super().__init__(
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
name, size_x, size_y, size_z, ordered_items=items, category=category, model=model, **kwargs
|
ordered_items=items,
|
||||||
)
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
elif ordering_param is not None:
|
elif ordering_param is not None:
|
||||||
# 传递 ordering 参数,让 TipRack 自己创建 Tip 对象
|
# 传递 ordering 参数,让 TipRack 自己创建 Tip 对象
|
||||||
super().__init__(
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
name, size_x, size_y, size_z, ordering=ordering_param, category=category, model=model, **kwargs
|
ordering=ordering_param,
|
||||||
)
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
else:
|
else:
|
||||||
super().__init__(name, size_x, size_y, size_z, category=category, model=model, **kwargs)
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
|
category=category,
|
||||||
|
model=model, **kwargs)
|
||||||
self._unilabos_state = {}
|
self._unilabos_state = {}
|
||||||
if material_info:
|
if material_info:
|
||||||
self._unilabos_state["Material"] = material_info
|
self._unilabos_state["Material"] = material_info
|
||||||
@@ -285,7 +255,7 @@ class PRCXI9300TipRack(TipRack):
|
|||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
data = {}
|
data = {}
|
||||||
if hasattr(self, "_unilabos_state") and self._unilabos_state:
|
if hasattr(self, '_unilabos_state') and self._unilabos_state:
|
||||||
safe_state = {}
|
safe_state = {}
|
||||||
for k, v in self._unilabos_state.items():
|
for k, v in self._unilabos_state.items():
|
||||||
# 如果是 Material 字典,深入检查
|
# 如果是 Material 字典,深入检查
|
||||||
@@ -307,23 +277,16 @@ class PRCXI9300TipRack(TipRack):
|
|||||||
data.update(safe_state)
|
data.update(safe_state)
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300Trash(Trash):
|
class PRCXI9300Trash(Trash):
|
||||||
"""PRCXI 9300 的专用 Trash 类,继承自 Trash。
|
"""PRCXI 9300 的专用 Trash 类,继承自 Trash。
|
||||||
|
|
||||||
该类定义了 PRCXI 9300 的工作台布局和槽位信息。
|
该类定义了 PRCXI 9300 的工作台布局和槽位信息。
|
||||||
"""
|
"""
|
||||||
|
|
||||||
def __init__(
|
def __init__(self, name: str, size_x: float, size_y: float, size_z: float,
|
||||||
self,
|
|
||||||
name: str,
|
|
||||||
size_x: float,
|
|
||||||
size_y: float,
|
|
||||||
size_z: float,
|
|
||||||
category: str = "trash",
|
category: str = "trash",
|
||||||
material_info: Optional[Dict[str, Any]] = None,
|
material_info: Optional[Dict[str, Any]] = None,
|
||||||
**kwargs,
|
**kwargs):
|
||||||
):
|
|
||||||
|
|
||||||
if name != "trash":
|
if name != "trash":
|
||||||
print(f"Warning: PRCXI9300Trash usually expects name='trash' for backend logic, but got '{name}'.")
|
print(f"Warning: PRCXI9300Trash usually expects name='trash' for backend logic, but got '{name}'.")
|
||||||
@@ -343,7 +306,7 @@ class PRCXI9300Trash(Trash):
|
|||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
data = {}
|
data = {}
|
||||||
if hasattr(self, "_unilabos_state") and self._unilabos_state:
|
if hasattr(self, '_unilabos_state') and self._unilabos_state:
|
||||||
safe_state = {}
|
safe_state = {}
|
||||||
for k, v in self._unilabos_state.items():
|
for k, v in self._unilabos_state.items():
|
||||||
# 如果是 Material 字典,深入检查
|
# 如果是 Material 字典,深入检查
|
||||||
@@ -365,27 +328,19 @@ class PRCXI9300Trash(Trash):
|
|||||||
data.update(safe_state)
|
data.update(safe_state)
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300TubeRack(TubeRack):
|
class PRCXI9300TubeRack(TubeRack):
|
||||||
"""
|
"""
|
||||||
专用管架类:用于 EP 管架、试管架等。
|
专用管架类:用于 EP 管架、试管架等。
|
||||||
继承自 PLR 的 TubeRack,并支持注入 material_info (UUID)。
|
继承自 PLR 的 TubeRack,并支持注入 material_info (UUID)。
|
||||||
"""
|
"""
|
||||||
|
def __init__(self, name: str, size_x: float, size_y: float, size_z: float,
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
name: str,
|
|
||||||
size_x: float,
|
|
||||||
size_y: float,
|
|
||||||
size_z: float,
|
|
||||||
category: str = "tube_rack",
|
category: str = "tube_rack",
|
||||||
items: Optional[Dict[str, Any]] = None,
|
items: Optional[Dict[str, Any]] = None,
|
||||||
ordered_items: Optional[OrderedDict] = None,
|
ordered_items: Optional[OrderedDict] = None,
|
||||||
ordering: Optional[OrderedDict] = None,
|
ordering: Optional[OrderedDict] = None,
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
material_info: Optional[Dict[str, Any]] = None,
|
material_info: Optional[Dict[str, Any]] = None,
|
||||||
**kwargs,
|
**kwargs):
|
||||||
):
|
|
||||||
|
|
||||||
# 如果 ordered_items 不为 None,直接使用
|
# 如果 ordered_items 不为 None,直接使用
|
||||||
if ordered_items is not None:
|
if ordered_items is not None:
|
||||||
@@ -415,12 +370,20 @@ class PRCXI9300TubeRack(TubeRack):
|
|||||||
|
|
||||||
# 根据情况传递不同的参数
|
# 根据情况传递不同的参数
|
||||||
if items_to_pass is not None:
|
if items_to_pass is not None:
|
||||||
super().__init__(name, size_x, size_y, size_z, ordered_items=items_to_pass, model=model, **kwargs)
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
|
ordered_items=items_to_pass,
|
||||||
|
model=model,
|
||||||
|
**kwargs)
|
||||||
elif ordering_param is not None:
|
elif ordering_param is not None:
|
||||||
# 传递 ordering 参数,让 TubeRack 自己创建 Tube 对象
|
# 传递 ordering 参数,让 TubeRack 自己创建 Tube 对象
|
||||||
super().__init__(name, size_x, size_y, size_z, ordering=ordering_param, model=model, **kwargs)
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
|
ordering=ordering_param,
|
||||||
|
model=model,
|
||||||
|
**kwargs)
|
||||||
else:
|
else:
|
||||||
super().__init__(name, size_x, size_y, size_z, model=model, **kwargs)
|
super().__init__(name, size_x, size_y, size_z,
|
||||||
|
model=model,
|
||||||
|
**kwargs)
|
||||||
|
|
||||||
self._unilabos_state = {}
|
self._unilabos_state = {}
|
||||||
if material_info:
|
if material_info:
|
||||||
@@ -431,7 +394,7 @@ class PRCXI9300TubeRack(TubeRack):
|
|||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
data = {}
|
data = {}
|
||||||
if hasattr(self, "_unilabos_state") and self._unilabos_state:
|
if hasattr(self, '_unilabos_state') and self._unilabos_state:
|
||||||
safe_state = {}
|
safe_state = {}
|
||||||
for k, v in self._unilabos_state.items():
|
for k, v in self._unilabos_state.items():
|
||||||
# 如果是 Material 字典,深入检查
|
# 如果是 Material 字典,深入检查
|
||||||
@@ -453,19 +416,12 @@ class PRCXI9300TubeRack(TubeRack):
|
|||||||
data.update(safe_state)
|
data.update(safe_state)
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300PlateAdapter(PlateAdapter):
|
class PRCXI9300PlateAdapter(PlateAdapter):
|
||||||
"""
|
"""
|
||||||
专用板式适配器类:用于承载 Plate 的底座(如 PCR 适配器、磁吸架等)。
|
专用板式适配器类:用于承载 Plate 的底座(如 PCR 适配器、磁吸架等)。
|
||||||
支持注入 material_info (UUID)。
|
支持注入 material_info (UUID)。
|
||||||
"""
|
"""
|
||||||
|
def __init__(self, name: str, size_x: float, size_y: float, size_z: float,
|
||||||
def __init__(
|
|
||||||
self,
|
|
||||||
name: str,
|
|
||||||
size_x: float,
|
|
||||||
size_y: float,
|
|
||||||
size_z: float,
|
|
||||||
category: str = "plate_adapter",
|
category: str = "plate_adapter",
|
||||||
model: Optional[str] = None,
|
model: Optional[str] = None,
|
||||||
material_info: Optional[Dict[str, Any]] = None,
|
material_info: Optional[Dict[str, Any]] = None,
|
||||||
@@ -476,8 +432,7 @@ class PRCXI9300PlateAdapter(PlateAdapter):
|
|||||||
dx: Optional[float] = None,
|
dx: Optional[float] = None,
|
||||||
dy: Optional[float] = None,
|
dy: Optional[float] = None,
|
||||||
dz: float = 0.0, # 默认Z轴偏移
|
dz: float = 0.0, # 默认Z轴偏移
|
||||||
**kwargs,
|
**kwargs):
|
||||||
):
|
|
||||||
|
|
||||||
# 自动居中计算:如果未指定 dx/dy,则根据适配器尺寸和孔尺寸计算居中位置
|
# 自动居中计算:如果未指定 dx/dy,则根据适配器尺寸和孔尺寸计算居中位置
|
||||||
if dx is None:
|
if dx is None:
|
||||||
@@ -497,7 +452,7 @@ class PRCXI9300PlateAdapter(PlateAdapter):
|
|||||||
adapter_hole_size_y=adapter_hole_size_y,
|
adapter_hole_size_y=adapter_hole_size_y,
|
||||||
adapter_hole_size_z=adapter_hole_size_z,
|
adapter_hole_size_z=adapter_hole_size_z,
|
||||||
model=model,
|
model=model,
|
||||||
**kwargs,
|
**kwargs
|
||||||
)
|
)
|
||||||
|
|
||||||
self._unilabos_state = {}
|
self._unilabos_state = {}
|
||||||
@@ -509,7 +464,7 @@ class PRCXI9300PlateAdapter(PlateAdapter):
|
|||||||
data = super().serialize_state()
|
data = super().serialize_state()
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
data = {}
|
data = {}
|
||||||
if hasattr(self, "_unilabos_state") and self._unilabos_state:
|
if hasattr(self, '_unilabos_state') and self._unilabos_state:
|
||||||
safe_state = {}
|
safe_state = {}
|
||||||
for k, v in self._unilabos_state.items():
|
for k, v in self._unilabos_state.items():
|
||||||
# 如果是 Material 字典,深入检查
|
# 如果是 Material 字典,深入检查
|
||||||
@@ -531,7 +486,6 @@ class PRCXI9300PlateAdapter(PlateAdapter):
|
|||||||
data.update(safe_state)
|
data.update(safe_state)
|
||||||
return data
|
return data
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300Handler(LiquidHandlerAbstract):
|
class PRCXI9300Handler(LiquidHandlerAbstract):
|
||||||
support_touch_tip = False
|
support_touch_tip = False
|
||||||
|
|
||||||
@@ -564,9 +518,7 @@ class PRCXI9300Handler(LiquidHandlerAbstract):
|
|||||||
if "Material" in child.children[0]._unilabos_state:
|
if "Material" in child.children[0]._unilabos_state:
|
||||||
number = int(child.name.replace("T", ""))
|
number = int(child.name.replace("T", ""))
|
||||||
tablets_info.append(
|
tablets_info.append(
|
||||||
WorkTablets(
|
WorkTablets(Number=number, Code=f"T{number}", Material=child.children[0]._unilabos_state["Material"])
|
||||||
Number=number, Code=f"T{number}", Material=child.children[0]._unilabos_state["Material"]
|
|
||||||
)
|
|
||||||
)
|
)
|
||||||
if is_9320:
|
if is_9320:
|
||||||
print("当前设备是9320")
|
print("当前设备是9320")
|
||||||
@@ -586,14 +538,9 @@ class PRCXI9300Handler(LiquidHandlerAbstract):
|
|||||||
super().post_init(ros_node)
|
super().post_init(ros_node)
|
||||||
self._unilabos_backend.post_init(ros_node)
|
self._unilabos_backend.post_init(ros_node)
|
||||||
|
|
||||||
def set_liquid(self, wells: list[Well], liquid_names: list[str], volumes: list[float]) -> SetLiquidReturn:
|
def set_liquid(self, wells: list[Well], liquid_names: list[str], volumes: list[float]) -> SimpleReturn:
|
||||||
return super().set_liquid(wells, liquid_names, volumes)
|
return super().set_liquid(wells, liquid_names, volumes)
|
||||||
|
|
||||||
def set_liquid_from_plate(
|
|
||||||
self, plate: ResourceSlot, well_names: list[str], liquid_names: list[str], volumes: list[float]
|
|
||||||
) -> SetLiquidFromPlateReturn:
|
|
||||||
return super().set_liquid_from_plate(plate, well_names, liquid_names, volumes)
|
|
||||||
|
|
||||||
def set_group(self, group_name: str, wells: List[Well], volumes: List[float]):
|
def set_group(self, group_name: str, wells: List[Well], volumes: List[float]):
|
||||||
return super().set_group(group_name, wells, volumes)
|
return super().set_group(group_name, wells, volumes)
|
||||||
|
|
||||||
@@ -853,7 +800,6 @@ class PRCXI9300Handler(LiquidHandlerAbstract):
|
|||||||
|
|
||||||
async def heater_action(self, temperature: float, time: int):
|
async def heater_action(self, temperature: float, time: int):
|
||||||
return await self._unilabos_backend.heater_action(temperature, time)
|
return await self._unilabos_backend.heater_action(temperature, time)
|
||||||
|
|
||||||
async def move_plate(
|
async def move_plate(
|
||||||
self,
|
self,
|
||||||
plate: Plate,
|
plate: Plate,
|
||||||
@@ -876,11 +822,10 @@ class PRCXI9300Handler(LiquidHandlerAbstract):
|
|||||||
drop_direction,
|
drop_direction,
|
||||||
pickup_direction,
|
pickup_direction,
|
||||||
pickup_distance_from_top,
|
pickup_distance_from_top,
|
||||||
target_plate_number=to,
|
target_plate_number = to,
|
||||||
**backend_kwargs,
|
**backend_kwargs,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class PRCXI9300Backend(LiquidHandlerBackend):
|
class PRCXI9300Backend(LiquidHandlerBackend):
|
||||||
"""PRCXI 9300 的后端实现,继承自 LiquidHandlerBackend。
|
"""PRCXI 9300 的后端实现,继承自 LiquidHandlerBackend。
|
||||||
|
|
||||||
@@ -933,12 +878,13 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
self.steps_todo_list.append(step)
|
self.steps_todo_list.append(step)
|
||||||
return step
|
return step
|
||||||
|
|
||||||
|
|
||||||
async def pick_up_resource(self, pickup: ResourcePickup, **backend_kwargs):
|
async def pick_up_resource(self, pickup: ResourcePickup, **backend_kwargs):
|
||||||
|
|
||||||
resource = pickup.resource
|
resource=pickup.resource
|
||||||
offset = pickup.offset
|
offset=pickup.offset
|
||||||
pickup_distance_from_top = pickup.pickup_distance_from_top
|
pickup_distance_from_top=pickup.pickup_distance_from_top
|
||||||
direction = pickup.direction
|
direction=pickup.direction
|
||||||
|
|
||||||
plate_number = int(resource.parent.name.replace("T", ""))
|
plate_number = int(resource.parent.name.replace("T", ""))
|
||||||
is_whole_plate = True
|
is_whole_plate = True
|
||||||
@@ -950,11 +896,13 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
|
|
||||||
async def drop_resource(self, drop: ResourceDrop, **backend_kwargs):
|
async def drop_resource(self, drop: ResourceDrop, **backend_kwargs):
|
||||||
|
|
||||||
|
|
||||||
plate_number = None
|
plate_number = None
|
||||||
target_plate_number = backend_kwargs.get("target_plate_number", None)
|
target_plate_number = backend_kwargs.get("target_plate_number", None)
|
||||||
if target_plate_number is not None:
|
if target_plate_number is not None:
|
||||||
plate_number = int(target_plate_number.name.replace("T", ""))
|
plate_number = int(target_plate_number.name.replace("T", ""))
|
||||||
|
|
||||||
|
|
||||||
is_whole_plate = True
|
is_whole_plate = True
|
||||||
balance_height = 0
|
balance_height = 0
|
||||||
if plate_number is None:
|
if plate_number is None:
|
||||||
@@ -963,6 +911,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
self.steps_todo_list.append(step)
|
self.steps_todo_list.append(step)
|
||||||
return step
|
return step
|
||||||
|
|
||||||
|
|
||||||
async def heater_action(self, temperature: float, time: int):
|
async def heater_action(self, temperature: float, time: int):
|
||||||
print(f"\n\nHeater action: temperature={temperature}, time={time}\n\n")
|
print(f"\n\nHeater action: temperature={temperature}, time={time}\n\n")
|
||||||
# return await self.api_client.heater_action(temperature, time)
|
# return await self.api_client.heater_action(temperature, time)
|
||||||
@@ -1031,7 +980,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
# 检查重置状态并等待完成
|
# 检查重置状态并等待完成
|
||||||
while not self.is_reset_ok:
|
while not self.is_reset_ok:
|
||||||
print("Waiting for PRCXI9300 to reset...")
|
print("Waiting for PRCXI9300 to reset...")
|
||||||
if hasattr(self, "_ros_node") and self._ros_node is not None:
|
if hasattr(self, '_ros_node') and self._ros_node is not None:
|
||||||
await self._ros_node.sleep(1)
|
await self._ros_node.sleep(1)
|
||||||
else:
|
else:
|
||||||
await asyncio.sleep(1)
|
await asyncio.sleep(1)
|
||||||
@@ -1049,7 +998,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
"""Pick up tips from the specified resource."""
|
"""Pick up tips from the specified resource."""
|
||||||
# INSERT_YOUR_CODE
|
# INSERT_YOUR_CODE
|
||||||
# Ensure use_channels is converted to a list of ints if it's an array
|
# Ensure use_channels is converted to a list of ints if it's an array
|
||||||
if hasattr(use_channels, "tolist"):
|
if hasattr(use_channels, 'tolist'):
|
||||||
_use_channels = use_channels.tolist()
|
_use_channels = use_channels.tolist()
|
||||||
else:
|
else:
|
||||||
_use_channels = list(use_channels) if use_channels is not None else None
|
_use_channels = list(use_channels) if use_channels is not None else None
|
||||||
@@ -1103,7 +1052,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
|
|
||||||
async def drop_tips(self, ops: List[Drop], use_channels: List[int] = None):
|
async def drop_tips(self, ops: List[Drop], use_channels: List[int] = None):
|
||||||
"""Pick up tips from the specified resource."""
|
"""Pick up tips from the specified resource."""
|
||||||
if hasattr(use_channels, "tolist"):
|
if hasattr(use_channels, 'tolist'):
|
||||||
_use_channels = use_channels.tolist()
|
_use_channels = use_channels.tolist()
|
||||||
else:
|
else:
|
||||||
_use_channels = list(use_channels) if use_channels is not None else None
|
_use_channels = list(use_channels) if use_channels is not None else None
|
||||||
@@ -1229,7 +1178,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
|
|
||||||
async def aspirate(self, ops: List[SingleChannelAspiration], use_channels: List[int] = None):
|
async def aspirate(self, ops: List[SingleChannelAspiration], use_channels: List[int] = None):
|
||||||
"""Aspirate liquid from the specified resources."""
|
"""Aspirate liquid from the specified resources."""
|
||||||
if hasattr(use_channels, "tolist"):
|
if hasattr(use_channels, 'tolist'):
|
||||||
_use_channels = use_channels.tolist()
|
_use_channels = use_channels.tolist()
|
||||||
else:
|
else:
|
||||||
_use_channels = list(use_channels) if use_channels is not None else None
|
_use_channels = list(use_channels) if use_channels is not None else None
|
||||||
@@ -1286,7 +1235,7 @@ class PRCXI9300Backend(LiquidHandlerBackend):
|
|||||||
|
|
||||||
async def dispense(self, ops: List[SingleChannelDispense], use_channels: List[int] = None):
|
async def dispense(self, ops: List[SingleChannelDispense], use_channels: List[int] = None):
|
||||||
"""Dispense liquid into the specified resources."""
|
"""Dispense liquid into the specified resources."""
|
||||||
if hasattr(use_channels, "tolist"):
|
if hasattr(use_channels, 'tolist'):
|
||||||
_use_channels = use_channels.tolist()
|
_use_channels = use_channels.tolist()
|
||||||
else:
|
else:
|
||||||
_use_channels = list(use_channels) if use_channels is not None else None
|
_use_channels = list(use_channels) if use_channels is not None else None
|
||||||
@@ -1467,6 +1416,7 @@ class PRCXI9300Api:
|
|||||||
time.sleep(1)
|
time.sleep(1)
|
||||||
return success
|
return success
|
||||||
|
|
||||||
|
|
||||||
def call(self, service: str, method: str, params: Optional[list] = None) -> Any:
|
def call(self, service: str, method: str, params: Optional[list] = None) -> Any:
|
||||||
payload = json.dumps(
|
payload = json.dumps(
|
||||||
{"ServiceName": service, "MethodName": method, "Paramters": params or []}, separators=(",", ":")
|
{"ServiceName": service, "MethodName": method, "Paramters": params or []}, separators=(",", ":")
|
||||||
@@ -1731,11 +1681,11 @@ class PRCXI9300Api:
|
|||||||
"LiquidDispensingMethod": liquid_method,
|
"LiquidDispensingMethod": liquid_method,
|
||||||
}
|
}
|
||||||
|
|
||||||
def clamp_jaw_pick_up(
|
def clamp_jaw_pick_up(self,
|
||||||
self,
|
|
||||||
plate_no: int,
|
plate_no: int,
|
||||||
is_whole_plate: bool,
|
is_whole_plate: bool,
|
||||||
balance_height: int,
|
balance_height: int,
|
||||||
|
|
||||||
) -> Dict[str, Any]:
|
) -> Dict[str, Any]:
|
||||||
return {
|
return {
|
||||||
"StepAxis": "ClampingJaw",
|
"StepAxis": "ClampingJaw",
|
||||||
@@ -1745,7 +1695,7 @@ class PRCXI9300Api:
|
|||||||
"HoleRow": 1,
|
"HoleRow": 1,
|
||||||
"HoleCol": 1,
|
"HoleCol": 1,
|
||||||
"BalanceHeight": balance_height,
|
"BalanceHeight": balance_height,
|
||||||
"PlateOrHoleNum": f"T{plate_no}",
|
"PlateOrHoleNum": f"T{plate_no}"
|
||||||
}
|
}
|
||||||
|
|
||||||
def clamp_jaw_drop(
|
def clamp_jaw_drop(
|
||||||
@@ -1753,6 +1703,7 @@ class PRCXI9300Api:
|
|||||||
plate_no: int,
|
plate_no: int,
|
||||||
is_whole_plate: bool,
|
is_whole_plate: bool,
|
||||||
balance_height: int,
|
balance_height: int,
|
||||||
|
|
||||||
) -> Dict[str, Any]:
|
) -> Dict[str, Any]:
|
||||||
return {
|
return {
|
||||||
"StepAxis": "ClampingJaw",
|
"StepAxis": "ClampingJaw",
|
||||||
@@ -1762,7 +1713,7 @@ class PRCXI9300Api:
|
|||||||
"HoleRow": 1,
|
"HoleRow": 1,
|
||||||
"HoleCol": 1,
|
"HoleCol": 1,
|
||||||
"BalanceHeight": balance_height,
|
"BalanceHeight": balance_height,
|
||||||
"PlateOrHoleNum": f"T{plate_no}",
|
"PlateOrHoleNum": f"T{plate_no}"
|
||||||
}
|
}
|
||||||
|
|
||||||
def shaker_action(self, time: int, module_no: int, amplitude: int, is_wait: bool):
|
def shaker_action(self, time: int, module_no: int, amplitude: int, is_wait: bool):
|
||||||
@@ -1775,7 +1726,6 @@ class PRCXI9300Api:
|
|||||||
"AssistFun4": is_wait,
|
"AssistFun4": is_wait,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class DefaultLayout:
|
class DefaultLayout:
|
||||||
|
|
||||||
def __init__(self, product_name: str = "PRCXI9300"):
|
def __init__(self, product_name: str = "PRCXI9300"):
|
||||||
@@ -2154,9 +2104,7 @@ if __name__ == "__main__":
|
|||||||
size_y=50,
|
size_y=50,
|
||||||
size_z=10,
|
size_z=10,
|
||||||
category="tip_rack",
|
category="tip_rack",
|
||||||
ordered_items=collections.OrderedDict(
|
ordered_items=collections.OrderedDict({k: f"{child_prefix}_{k}" for k, v in tip_racks["ordering"].items()}),
|
||||||
{k: f"{child_prefix}_{k}" for k, v in tip_racks["ordering"].items()}
|
|
||||||
),
|
|
||||||
)
|
)
|
||||||
tip_rack_serialized = tip_rack.serialize()
|
tip_rack_serialized = tip_rack.serialize()
|
||||||
tip_rack_serialized["parent_name"] = deck.name
|
tip_rack_serialized["parent_name"] = deck.name
|
||||||
@@ -2351,19 +2299,21 @@ if __name__ == "__main__":
|
|||||||
|
|
||||||
A = tree_to_list([resource_plr_to_ulab(deck)])
|
A = tree_to_list([resource_plr_to_ulab(deck)])
|
||||||
with open("deck.json", "w", encoding="utf-8") as f:
|
with open("deck.json", "w", encoding="utf-8") as f:
|
||||||
A.insert(
|
A.insert(0, {
|
||||||
0,
|
|
||||||
{
|
|
||||||
"id": "PRCXI",
|
"id": "PRCXI",
|
||||||
"name": "PRCXI",
|
"name": "PRCXI",
|
||||||
"parent": None,
|
"parent": None,
|
||||||
"type": "device",
|
"type": "device",
|
||||||
"class": "liquid_handler.prcxi",
|
"class": "liquid_handler.prcxi",
|
||||||
"position": {"x": 0, "y": 0, "z": 0},
|
"position": {
|
||||||
|
"x": 0,
|
||||||
|
"y": 0,
|
||||||
|
"z": 0
|
||||||
|
},
|
||||||
"config": {
|
"config": {
|
||||||
"deck": {
|
"deck": {
|
||||||
"_resource_child_name": "PRCXI_Deck",
|
"_resource_child_name": "PRCXI_Deck",
|
||||||
"_resource_type": "unilabos.devices.liquid_handling.prcxi.prcxi:PRCXI9300Deck",
|
"_resource_type": "unilabos.devices.liquid_handling.prcxi.prcxi:PRCXI9300Deck"
|
||||||
},
|
},
|
||||||
"host": "192.168.0.121",
|
"host": "192.168.0.121",
|
||||||
"port": 9999,
|
"port": 9999,
|
||||||
@@ -2374,14 +2324,18 @@ if __name__ == "__main__":
|
|||||||
"debug": True,
|
"debug": True,
|
||||||
"simulator": True,
|
"simulator": True,
|
||||||
"matrix_id": "5de524d0-3f95-406c-86dd-f83626ebc7cb",
|
"matrix_id": "5de524d0-3f95-406c-86dd-f83626ebc7cb",
|
||||||
"is_9320": True,
|
"is_9320": True
|
||||||
},
|
},
|
||||||
"data": {},
|
"data": {},
|
||||||
"children": ["PRCXI_Deck"],
|
"children": [
|
||||||
},
|
"PRCXI_Deck"
|
||||||
)
|
]
|
||||||
|
})
|
||||||
A[1]["parent"] = "PRCXI"
|
A[1]["parent"] = "PRCXI"
|
||||||
json.dump({"nodes": A, "links": []}, f, indent=4, ensure_ascii=False)
|
json.dump({
|
||||||
|
"nodes": A,
|
||||||
|
"links": []
|
||||||
|
}, f, indent=4, ensure_ascii=False)
|
||||||
|
|
||||||
handler = PRCXI9300Handler(
|
handler = PRCXI9300Handler(
|
||||||
deck=deck,
|
deck=deck,
|
||||||
@@ -2423,6 +2377,7 @@ if __name__ == "__main__":
|
|||||||
time.sleep(5)
|
time.sleep(5)
|
||||||
os._exit(0)
|
os._exit(0)
|
||||||
|
|
||||||
|
|
||||||
prcxi_api = PRCXI9300Api(host="192.168.0.121", port=9999)
|
prcxi_api = PRCXI9300Api(host="192.168.0.121", port=9999)
|
||||||
prcxi_api.list_matrices()
|
prcxi_api.list_matrices()
|
||||||
prcxi_api.get_all_materials()
|
prcxi_api.get_all_materials()
|
||||||
|
|||||||
@@ -9284,13 +9284,7 @@ liquid_handler.prcxi:
|
|||||||
data_source: handle
|
data_source: handle
|
||||||
data_type: resource
|
data_type: resource
|
||||||
handler_key: input_wells
|
handler_key: input_wells
|
||||||
label: 待设定液体孔
|
label: InputWells
|
||||||
output:
|
|
||||||
- data_key: wells.@flatten
|
|
||||||
data_source: executor
|
|
||||||
data_type: resource
|
|
||||||
handler_key: output_wells
|
|
||||||
label: 已设定液体孔
|
|
||||||
placeholder_keys:
|
placeholder_keys:
|
||||||
wells: unilabos_resources
|
wells: unilabos_resources
|
||||||
result: {}
|
result: {}
|
||||||
@@ -9406,163 +9400,6 @@ liquid_handler.prcxi:
|
|||||||
title: LiquidHandlerSetLiquid
|
title: LiquidHandlerSetLiquid
|
||||||
type: object
|
type: object
|
||||||
type: LiquidHandlerSetLiquid
|
type: LiquidHandlerSetLiquid
|
||||||
set_liquid_from_plate:
|
|
||||||
feedback: {}
|
|
||||||
goal: {}
|
|
||||||
goal_default:
|
|
||||||
liquid_names: null
|
|
||||||
plate: null
|
|
||||||
volumes: null
|
|
||||||
well_names: null
|
|
||||||
handles:
|
|
||||||
input:
|
|
||||||
- data_key: plate
|
|
||||||
data_source: handle
|
|
||||||
data_type: resource
|
|
||||||
handler_key: input_plate
|
|
||||||
label: 待设定液体板
|
|
||||||
output:
|
|
||||||
- data_key: plate.@flatten
|
|
||||||
data_source: executor
|
|
||||||
data_type: resource
|
|
||||||
handler_key: output_plate
|
|
||||||
label: 已设定液体板
|
|
||||||
- data_key: wells.@flatten
|
|
||||||
data_source: executor
|
|
||||||
data_type: resource
|
|
||||||
handler_key: output_wells
|
|
||||||
label: 已设定液体孔
|
|
||||||
- data_key: volumes
|
|
||||||
data_source: executor
|
|
||||||
data_type: number_array
|
|
||||||
handler_key: output_volumes
|
|
||||||
label: 各孔设定体积
|
|
||||||
placeholder_keys:
|
|
||||||
plate: unilabos_resources
|
|
||||||
result: {}
|
|
||||||
schema:
|
|
||||||
description: ''
|
|
||||||
properties:
|
|
||||||
feedback: {}
|
|
||||||
goal:
|
|
||||||
properties:
|
|
||||||
liquid_names:
|
|
||||||
items:
|
|
||||||
type: string
|
|
||||||
type: array
|
|
||||||
plate:
|
|
||||||
properties:
|
|
||||||
category:
|
|
||||||
type: string
|
|
||||||
children:
|
|
||||||
items:
|
|
||||||
type: string
|
|
||||||
type: array
|
|
||||||
config:
|
|
||||||
type: string
|
|
||||||
data:
|
|
||||||
type: string
|
|
||||||
id:
|
|
||||||
type: string
|
|
||||||
name:
|
|
||||||
type: string
|
|
||||||
parent:
|
|
||||||
type: string
|
|
||||||
pose:
|
|
||||||
properties:
|
|
||||||
orientation:
|
|
||||||
properties:
|
|
||||||
w:
|
|
||||||
type: number
|
|
||||||
x:
|
|
||||||
type: number
|
|
||||||
y:
|
|
||||||
type: number
|
|
||||||
z:
|
|
||||||
type: number
|
|
||||||
required:
|
|
||||||
- x
|
|
||||||
- y
|
|
||||||
- z
|
|
||||||
- w
|
|
||||||
title: orientation
|
|
||||||
type: object
|
|
||||||
position:
|
|
||||||
properties:
|
|
||||||
x:
|
|
||||||
type: number
|
|
||||||
y:
|
|
||||||
type: number
|
|
||||||
z:
|
|
||||||
type: number
|
|
||||||
required:
|
|
||||||
- x
|
|
||||||
- y
|
|
||||||
- z
|
|
||||||
title: position
|
|
||||||
type: object
|
|
||||||
required:
|
|
||||||
- position
|
|
||||||
- orientation
|
|
||||||
title: pose
|
|
||||||
type: object
|
|
||||||
sample_id:
|
|
||||||
type: string
|
|
||||||
type:
|
|
||||||
type: string
|
|
||||||
required:
|
|
||||||
- id
|
|
||||||
- name
|
|
||||||
- sample_id
|
|
||||||
- children
|
|
||||||
- parent
|
|
||||||
- type
|
|
||||||
- category
|
|
||||||
- pose
|
|
||||||
- config
|
|
||||||
- data
|
|
||||||
title: plate
|
|
||||||
type: object
|
|
||||||
volumes:
|
|
||||||
items:
|
|
||||||
type: number
|
|
||||||
type: array
|
|
||||||
well_names:
|
|
||||||
items:
|
|
||||||
type: string
|
|
||||||
type: array
|
|
||||||
required:
|
|
||||||
- plate
|
|
||||||
- well_names
|
|
||||||
- liquid_names
|
|
||||||
- volumes
|
|
||||||
type: object
|
|
||||||
result:
|
|
||||||
properties:
|
|
||||||
plate:
|
|
||||||
items: {}
|
|
||||||
title: Plate
|
|
||||||
type: array
|
|
||||||
volumes:
|
|
||||||
items:
|
|
||||||
type: number
|
|
||||||
title: Volumes
|
|
||||||
type: array
|
|
||||||
wells:
|
|
||||||
items: {}
|
|
||||||
title: Wells
|
|
||||||
type: array
|
|
||||||
required:
|
|
||||||
- plate
|
|
||||||
- wells
|
|
||||||
- volumes
|
|
||||||
title: SetLiquidFromPlateReturn
|
|
||||||
type: object
|
|
||||||
required:
|
|
||||||
- goal
|
|
||||||
title: set_liquid_from_plate参数
|
|
||||||
type: object
|
|
||||||
type: UniLabJsonCommand
|
|
||||||
set_tiprack:
|
set_tiprack:
|
||||||
feedback: {}
|
feedback: {}
|
||||||
goal:
|
goal:
|
||||||
@@ -9908,21 +9745,21 @@ liquid_handler.prcxi:
|
|||||||
- 0
|
- 0
|
||||||
handles:
|
handles:
|
||||||
input:
|
input:
|
||||||
- data_key: sources
|
- data_key: liquid
|
||||||
data_source: handle
|
data_source: handle
|
||||||
data_type: resource
|
data_type: resource
|
||||||
handler_key: sources_identifier
|
handler_key: sources
|
||||||
label: 待移动液体
|
label: sources
|
||||||
- data_key: targets
|
- data_key: liquid
|
||||||
data_source: handle
|
data_source: executor
|
||||||
data_type: resource
|
data_type: resource
|
||||||
handler_key: targets_identifier
|
handler_key: targets
|
||||||
label: 转移目标
|
label: targets
|
||||||
- data_key: tip_rack
|
- data_key: liquid
|
||||||
data_source: handle
|
data_source: executor
|
||||||
data_type: resource
|
data_type: resource
|
||||||
handler_key: tip_rack_identifier
|
handler_key: tip_rack
|
||||||
label: 墙头盒
|
label: tip_rack
|
||||||
output:
|
output:
|
||||||
- data_key: liquid
|
- data_key: liquid
|
||||||
data_source: handle
|
data_source: handle
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -163,7 +163,6 @@ class Registry:
|
|||||||
"res_id": "unilabos_resources", # 将当前实验室的全部物料id作为下拉框可选择
|
"res_id": "unilabos_resources", # 将当前实验室的全部物料id作为下拉框可选择
|
||||||
"device_id": "unilabos_devices", # 将当前实验室的全部设备id作为下拉框可选择
|
"device_id": "unilabos_devices", # 将当前实验室的全部设备id作为下拉框可选择
|
||||||
"parent": "unilabos_nodes", # 将当前实验室的设备/物料作为下拉框可选择
|
"parent": "unilabos_nodes", # 将当前实验室的设备/物料作为下拉框可选择
|
||||||
"class_name": "unilabos_class",
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
"test_latency": {
|
"test_latency": {
|
||||||
@@ -266,7 +265,7 @@ class Registry:
|
|||||||
abs_path = Path(path).absolute()
|
abs_path = Path(path).absolute()
|
||||||
resource_path = abs_path / "resources"
|
resource_path = abs_path / "resources"
|
||||||
files = list(resource_path.glob("*/*.yaml"))
|
files = list(resource_path.glob("*/*.yaml"))
|
||||||
logger.debug(f"[UniLab Registry] resources: {resource_path.exists()}, total: {len(files)}")
|
logger.trace(f"[UniLab Registry] load resources? {resource_path.exists()}, total: {len(files)}")
|
||||||
current_resource_number = len(self.resource_type_registry) + 1
|
current_resource_number = len(self.resource_type_registry) + 1
|
||||||
for i, file in enumerate(files):
|
for i, file in enumerate(files):
|
||||||
with open(file, encoding="utf-8", mode="r") as f:
|
with open(file, encoding="utf-8", mode="r") as f:
|
||||||
|
|||||||
@@ -42,7 +42,7 @@ def canonicalize_nodes_data(
|
|||||||
Returns:
|
Returns:
|
||||||
ResourceTreeSet: 标准化后的资源树集合
|
ResourceTreeSet: 标准化后的资源树集合
|
||||||
"""
|
"""
|
||||||
print_status(f"{len(nodes)} Resources loaded:", "info")
|
print_status(f"{len(nodes)} Resources loaded", "info")
|
||||||
|
|
||||||
# 第一步:基本预处理(处理graphml的label字段)
|
# 第一步:基本预处理(处理graphml的label字段)
|
||||||
outer_host_node_id = None
|
outer_host_node_id = None
|
||||||
@@ -260,7 +260,7 @@ def read_node_link_json(
|
|||||||
resource_tree_set = canonicalize_nodes_data(nodes)
|
resource_tree_set = canonicalize_nodes_data(nodes)
|
||||||
|
|
||||||
# 标准化边数据
|
# 标准化边数据
|
||||||
links = data.get("links", data.get("edges", []))
|
links = data.get("links", [])
|
||||||
standardized_links = canonicalize_links_ports(links, resource_tree_set)
|
standardized_links = canonicalize_links_ports(links, resource_tree_set)
|
||||||
|
|
||||||
# 构建 NetworkX 图(需要转换回 dict 格式)
|
# 构建 NetworkX 图(需要转换回 dict 格式)
|
||||||
@@ -597,8 +597,6 @@ def resource_plr_to_ulab(resource_plr: "ResourcePLR", parent_name: str = None, w
|
|||||||
"tube": "tube",
|
"tube": "tube",
|
||||||
"bottle_carrier": "bottle_carrier",
|
"bottle_carrier": "bottle_carrier",
|
||||||
"plate_adapter": "plate_adapter",
|
"plate_adapter": "plate_adapter",
|
||||||
"electrode_sheet": "electrode_sheet",
|
|
||||||
"material_hole": "material_hole",
|
|
||||||
}
|
}
|
||||||
if source in replace_info:
|
if source in replace_info:
|
||||||
return replace_info[source]
|
return replace_info[source]
|
||||||
|
|||||||
@@ -13,9 +13,6 @@ if TYPE_CHECKING:
|
|||||||
from pylabrobot.resources import Resource as PLRResource
|
from pylabrobot.resources import Resource as PLRResource
|
||||||
|
|
||||||
|
|
||||||
EXTRA_CLASS = "unilabos_resource_class"
|
|
||||||
|
|
||||||
|
|
||||||
class ResourceDictPositionSize(BaseModel):
|
class ResourceDictPositionSize(BaseModel):
|
||||||
depth: float = Field(description="Depth", default=0.0) # z
|
depth: float = Field(description="Depth", default=0.0) # z
|
||||||
width: float = Field(description="Width", default=0.0) # x
|
width: float = Field(description="Width", default=0.0) # x
|
||||||
@@ -69,8 +66,8 @@ class ResourceDict(BaseModel):
|
|||||||
klass: str = Field(alias="class", description="Resource class name")
|
klass: str = Field(alias="class", description="Resource class name")
|
||||||
pose: ResourceDictPosition = Field(description="Resource position", default_factory=ResourceDictPosition)
|
pose: ResourceDictPosition = Field(description="Resource position", default_factory=ResourceDictPosition)
|
||||||
config: Dict[str, Any] = Field(description="Resource configuration")
|
config: Dict[str, Any] = Field(description="Resource configuration")
|
||||||
data: Dict[str, Any] = Field(description="Resource data")
|
data: Dict[str, Any] = Field(description="Resource data, eg: container liquid data")
|
||||||
extra: Dict[str, Any] = Field(description="Extra data")
|
extra: Dict[str, Any] = Field(description="Extra data, eg: slot index")
|
||||||
|
|
||||||
@field_serializer("parent_uuid")
|
@field_serializer("parent_uuid")
|
||||||
def _serialize_parent(self, parent_uuid: Optional["ResourceDict"]):
|
def _serialize_parent(self, parent_uuid: Optional["ResourceDict"]):
|
||||||
@@ -396,7 +393,7 @@ class ResourceTreeSet(object):
|
|||||||
"parent": parent_resource, # 直接传入 ResourceDict 对象
|
"parent": parent_resource, # 直接传入 ResourceDict 对象
|
||||||
"parent_uuid": parent_uuid, # 使用 parent_uuid 而不是 parent 对象
|
"parent_uuid": parent_uuid, # 使用 parent_uuid 而不是 parent 对象
|
||||||
"type": replace_plr_type(d.get("category", "")),
|
"type": replace_plr_type(d.get("category", "")),
|
||||||
"class": extra.get(EXTRA_CLASS, ""),
|
"class": d.get("class", ""),
|
||||||
"position": pos,
|
"position": pos,
|
||||||
"pose": pos,
|
"pose": pos,
|
||||||
"config": {
|
"config": {
|
||||||
@@ -446,7 +443,7 @@ class ResourceTreeSet(object):
|
|||||||
trees.append(tree_instance)
|
trees.append(tree_instance)
|
||||||
return cls(trees)
|
return cls(trees)
|
||||||
|
|
||||||
def to_plr_resources(self, skip_devices=True) -> List["PLRResource"]:
|
def to_plr_resources(self) -> List["PLRResource"]:
|
||||||
"""
|
"""
|
||||||
将 ResourceTreeSet 转换为 PLR 资源列表
|
将 ResourceTreeSet 转换为 PLR 资源列表
|
||||||
|
|
||||||
@@ -471,7 +468,6 @@ class ResourceTreeSet(object):
|
|||||||
name_to_uuid[node.res_content.name] = node.res_content.uuid
|
name_to_uuid[node.res_content.name] = node.res_content.uuid
|
||||||
all_states[node.res_content.name] = node.res_content.data
|
all_states[node.res_content.name] = node.res_content.data
|
||||||
name_to_extra[node.res_content.name] = node.res_content.extra
|
name_to_extra[node.res_content.name] = node.res_content.extra
|
||||||
name_to_extra[node.res_content.name][EXTRA_CLASS] = node.res_content.klass
|
|
||||||
for child in node.children:
|
for child in node.children:
|
||||||
collect_node_data(child, name_to_uuid, all_states, name_to_extra)
|
collect_node_data(child, name_to_uuid, all_states, name_to_extra)
|
||||||
|
|
||||||
@@ -516,10 +512,7 @@ class ResourceTreeSet(object):
|
|||||||
plr_dict = node_to_plr_dict(tree.root_node, has_model)
|
plr_dict = node_to_plr_dict(tree.root_node, has_model)
|
||||||
try:
|
try:
|
||||||
sub_cls = find_subclass(plr_dict["type"], PLRResource)
|
sub_cls = find_subclass(plr_dict["type"], PLRResource)
|
||||||
if skip_devices and plr_dict["type"] == "device":
|
if sub_cls is None:
|
||||||
logger.info(f"跳过更新 {plr_dict['name']} 设备是class")
|
|
||||||
continue
|
|
||||||
elif sub_cls is None:
|
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
f"无法找到类型 {plr_dict['type']} 对应的 PLR 资源类。原始信息:{tree.root_node.res_content}"
|
f"无法找到类型 {plr_dict['type']} 对应的 PLR 资源类。原始信息:{tree.root_node.res_content}"
|
||||||
)
|
)
|
||||||
@@ -527,10 +520,6 @@ class ResourceTreeSet(object):
|
|||||||
if "category" not in spec.parameters:
|
if "category" not in spec.parameters:
|
||||||
plr_dict.pop("category", None)
|
plr_dict.pop("category", None)
|
||||||
plr_resource = sub_cls.deserialize(plr_dict, allow_marshal=True)
|
plr_resource = sub_cls.deserialize(plr_dict, allow_marshal=True)
|
||||||
from pylabrobot.resources import Coordinate
|
|
||||||
from pylabrobot.serializer import deserialize
|
|
||||||
location = cast(Coordinate, deserialize(plr_dict["location"]))
|
|
||||||
plr_resource.location = location
|
|
||||||
plr_resource.load_all_state(all_states)
|
plr_resource.load_all_state(all_states)
|
||||||
# 使用 DeviceNodeResourceTracker 设置 UUID 和 Extra
|
# 使用 DeviceNodeResourceTracker 设置 UUID 和 Extra
|
||||||
tracker.loop_set_uuid(plr_resource, name_to_uuid)
|
tracker.loop_set_uuid(plr_resource, name_to_uuid)
|
||||||
@@ -997,7 +986,7 @@ class DeviceNodeResourceTracker(object):
|
|||||||
extra = name_to_extra_map[resource_name]
|
extra = name_to_extra_map[resource_name]
|
||||||
self.set_resource_extra(res, extra)
|
self.set_resource_extra(res, extra)
|
||||||
if len(extra):
|
if len(extra):
|
||||||
logger.trace(f"设置资源Extra: {resource_name} -> {extra}")
|
logger.debug(f"设置资源Extra: {resource_name} -> {extra}")
|
||||||
return 1
|
return 1
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
|
|||||||
@@ -49,6 +49,7 @@ from unilabos.resources.resource_tracker import (
|
|||||||
ResourceTreeInstance,
|
ResourceTreeInstance,
|
||||||
ResourceDictInstance,
|
ResourceDictInstance,
|
||||||
)
|
)
|
||||||
|
from unilabos.ros.x.rclpyx import get_event_loop
|
||||||
from unilabos.ros.utils.driver_creator import WorkstationNodeCreator, PyLabRobotCreator, DeviceClassCreator
|
from unilabos.ros.utils.driver_creator import WorkstationNodeCreator, PyLabRobotCreator, DeviceClassCreator
|
||||||
from rclpy.task import Task, Future
|
from rclpy.task import Task, Future
|
||||||
from unilabos.utils.import_manager import default_manager
|
from unilabos.utils.import_manager import default_manager
|
||||||
@@ -184,7 +185,7 @@ class PropertyPublisher:
|
|||||||
f"创建发布者 {name} 失败,可能由于注册表有误,类型: {msg_type},错误: {ex}\n{traceback.format_exc()}"
|
f"创建发布者 {name} 失败,可能由于注册表有误,类型: {msg_type},错误: {ex}\n{traceback.format_exc()}"
|
||||||
)
|
)
|
||||||
self.timer = node.create_timer(self.timer_period, self.publish_property)
|
self.timer = node.create_timer(self.timer_period, self.publish_property)
|
||||||
self.__loop = ROS2DeviceNode.get_asyncio_loop()
|
self.__loop = get_event_loop()
|
||||||
str_msg_type = str(msg_type)[8:-2]
|
str_msg_type = str(msg_type)[8:-2]
|
||||||
self.node.lab_logger().trace(f"发布属性: {name}, 类型: {str_msg_type}, 周期: {initial_period}秒, QoS: {qos}")
|
self.node.lab_logger().trace(f"发布属性: {name}, 类型: {str_msg_type}, 周期: {initial_period}秒, QoS: {qos}")
|
||||||
|
|
||||||
@@ -884,9 +885,6 @@ class BaseROS2DeviceNode(Node, Generic[T]):
|
|||||||
parent_appended = True
|
parent_appended = True
|
||||||
|
|
||||||
# 加载状态
|
# 加载状态
|
||||||
original_instance.location = plr_resource.location
|
|
||||||
original_instance.rotation = plr_resource.rotation
|
|
||||||
original_instance.barcode = plr_resource.barcode
|
|
||||||
original_instance.load_all_state(states)
|
original_instance.load_all_state(states)
|
||||||
child_count = len(original_instance.get_all_children())
|
child_count = len(original_instance.get_all_children())
|
||||||
self.lab_logger().info(
|
self.lab_logger().info(
|
||||||
@@ -1322,32 +1320,19 @@ class BaseROS2DeviceNode(Node, Generic[T]):
|
|||||||
resource_inputs = action_kwargs[k] if is_sequence else [action_kwargs[k]]
|
resource_inputs = action_kwargs[k] if is_sequence else [action_kwargs[k]]
|
||||||
|
|
||||||
# 批量查询资源
|
# 批量查询资源
|
||||||
queried_resources: list = [None] * len(resource_inputs)
|
queried_resources = []
|
||||||
uuid_indices: list[tuple[int, str, dict]] = [] # (index, uuid, resource_data)
|
for resource_data in resource_inputs:
|
||||||
|
|
||||||
# 第一遍:处理没有uuid的资源,收集有uuid的资源信息
|
|
||||||
for idx, resource_data in enumerate(resource_inputs):
|
|
||||||
unilabos_uuid = resource_data.get("data", {}).get("unilabos_uuid")
|
unilabos_uuid = resource_data.get("data", {}).get("unilabos_uuid")
|
||||||
if unilabos_uuid is None:
|
if unilabos_uuid is None:
|
||||||
plr_resource = await self.get_resource_with_dir(
|
plr_resource = await self.get_resource_with_dir(
|
||||||
resource_id=resource_data["id"], with_children=True
|
resource_id=resource_data["id"], with_children=True
|
||||||
)
|
)
|
||||||
if "sample_id" in resource_data:
|
|
||||||
plr_resource.unilabos_extra["sample_uuid"] = resource_data["sample_id"]
|
|
||||||
queried_resources[idx] = plr_resource
|
|
||||||
else:
|
else:
|
||||||
uuid_indices.append((idx, unilabos_uuid, resource_data))
|
resource_tree = await self.get_resource([unilabos_uuid])
|
||||||
|
plr_resource = resource_tree.to_plr_resources()[0]
|
||||||
# 第二遍:批量查询有uuid的资源
|
|
||||||
if uuid_indices:
|
|
||||||
uuids = [item[1] for item in uuid_indices]
|
|
||||||
resource_tree = await self.get_resource(uuids)
|
|
||||||
plr_resources = resource_tree.to_plr_resources()
|
|
||||||
for i, (idx, _, resource_data) in enumerate(uuid_indices):
|
|
||||||
plr_resource = plr_resources[i]
|
|
||||||
if "sample_id" in resource_data:
|
if "sample_id" in resource_data:
|
||||||
plr_resource.unilabos_extra["sample_uuid"] = resource_data["sample_id"]
|
plr_resource.unilabos_extra["sample_uuid"] = resource_data["sample_id"]
|
||||||
queried_resources[idx] = plr_resource
|
queried_resources.append(plr_resource)
|
||||||
|
|
||||||
self.lab_logger().debug(f"资源查询结果: 共 {len(queried_resources)} 个资源")
|
self.lab_logger().debug(f"资源查询结果: 共 {len(queried_resources)} 个资源")
|
||||||
|
|
||||||
@@ -1772,15 +1757,6 @@ class ROS2DeviceNode:
|
|||||||
它不继承设备类,而是通过代理模式访问设备类的属性和方法。
|
它不继承设备类,而是通过代理模式访问设备类的属性和方法。
|
||||||
"""
|
"""
|
||||||
|
|
||||||
# 类变量,用于循环管理
|
|
||||||
_asyncio_loop = None
|
|
||||||
_asyncio_loop_running = False
|
|
||||||
_asyncio_loop_thread = None
|
|
||||||
|
|
||||||
@classmethod
|
|
||||||
def get_asyncio_loop(cls):
|
|
||||||
return cls._asyncio_loop
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
async def safe_task_wrapper(trace_callback, func, **kwargs):
|
async def safe_task_wrapper(trace_callback, func, **kwargs):
|
||||||
try:
|
try:
|
||||||
@@ -1857,11 +1833,6 @@ class ROS2DeviceNode:
|
|||||||
print_publish: 是否打印发布信息
|
print_publish: 是否打印发布信息
|
||||||
driver_is_ros:
|
driver_is_ros:
|
||||||
"""
|
"""
|
||||||
# 在初始化时检查循环状态
|
|
||||||
if ROS2DeviceNode._asyncio_loop_running and ROS2DeviceNode._asyncio_loop_thread is not None:
|
|
||||||
pass
|
|
||||||
elif ROS2DeviceNode._asyncio_loop_thread is None:
|
|
||||||
self._start_loop()
|
|
||||||
|
|
||||||
# 保存设备类是否支持异步上下文
|
# 保存设备类是否支持异步上下文
|
||||||
self._has_async_context = hasattr(driver_class, "__aenter__") and hasattr(driver_class, "__aexit__")
|
self._has_async_context = hasattr(driver_class, "__aenter__") and hasattr(driver_class, "__aexit__")
|
||||||
@@ -1953,17 +1924,6 @@ class ROS2DeviceNode:
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
self._ros_node.lab_logger().error(f"设备后初始化失败: {e}")
|
self._ros_node.lab_logger().error(f"设备后初始化失败: {e}")
|
||||||
|
|
||||||
def _start_loop(self):
|
|
||||||
def run_event_loop():
|
|
||||||
loop = asyncio.new_event_loop()
|
|
||||||
ROS2DeviceNode._asyncio_loop = loop
|
|
||||||
asyncio.set_event_loop(loop)
|
|
||||||
loop.run_forever()
|
|
||||||
|
|
||||||
ROS2DeviceNode._asyncio_loop_thread = threading.Thread(target=run_event_loop, daemon=True, name="ROS2DeviceNode")
|
|
||||||
ROS2DeviceNode._asyncio_loop_thread.start()
|
|
||||||
logger.info(f"循环线程已启动")
|
|
||||||
|
|
||||||
|
|
||||||
class DeviceInfoType(TypedDict):
|
class DeviceInfoType(TypedDict):
|
||||||
id: str
|
id: str
|
||||||
|
|||||||
@@ -808,7 +808,6 @@ class HostNode(BaseROS2DeviceNode):
|
|||||||
goal_msg = convert_to_ros_msg(action_client._action_type.Goal(), action_kwargs)
|
goal_msg = convert_to_ros_msg(action_client._action_type.Goal(), action_kwargs)
|
||||||
|
|
||||||
self.lab_logger().info(f"[Host Node] Sending goal for {action_id}: {str(goal_msg)[:1000]}")
|
self.lab_logger().info(f"[Host Node] Sending goal for {action_id}: {str(goal_msg)[:1000]}")
|
||||||
self.lab_logger().trace(f"[Host Node] Sending goal for {action_id}: {action_kwargs}")
|
|
||||||
self.lab_logger().trace(f"[Host Node] Sending goal for {action_id}: {goal_msg}")
|
self.lab_logger().trace(f"[Host Node] Sending goal for {action_id}: {goal_msg}")
|
||||||
action_client.wait_for_server()
|
action_client.wait_for_server()
|
||||||
goal_uuid_obj = UUID(uuid=list(u.bytes))
|
goal_uuid_obj = UUID(uuid=list(u.bytes))
|
||||||
|
|||||||
@@ -6,6 +6,8 @@ from typing import List, Dict, Any, Optional, TYPE_CHECKING
|
|||||||
|
|
||||||
import rclpy
|
import rclpy
|
||||||
from rosidl_runtime_py import message_to_ordereddict
|
from rosidl_runtime_py import message_to_ordereddict
|
||||||
|
from unilabos_msgs.msg import Resource
|
||||||
|
from unilabos_msgs.srv import ResourceUpdate
|
||||||
|
|
||||||
from unilabos.messages import * # type: ignore # protocol names
|
from unilabos.messages import * # type: ignore # protocol names
|
||||||
from rclpy.action import ActionServer, ActionClient
|
from rclpy.action import ActionServer, ActionClient
|
||||||
@@ -13,6 +15,7 @@ from rclpy.action.server import ServerGoalHandle
|
|||||||
from unilabos_msgs.srv._serial_command import SerialCommand_Request, SerialCommand_Response
|
from unilabos_msgs.srv._serial_command import SerialCommand_Request, SerialCommand_Response
|
||||||
|
|
||||||
from unilabos.compile import action_protocol_generators
|
from unilabos.compile import action_protocol_generators
|
||||||
|
from unilabos.resources.graphio import nested_dict_to_list
|
||||||
from unilabos.ros.initialize_device import initialize_device_from_dict
|
from unilabos.ros.initialize_device import initialize_device_from_dict
|
||||||
from unilabos.ros.msgs.message_converter import (
|
from unilabos.ros.msgs.message_converter import (
|
||||||
get_action_type,
|
get_action_type,
|
||||||
@@ -228,15 +231,15 @@ class ROS2WorkstationNode(BaseROS2DeviceNode):
|
|||||||
try:
|
try:
|
||||||
# 统一处理单个或多个资源
|
# 统一处理单个或多个资源
|
||||||
resource_id = (
|
resource_id = (
|
||||||
protocol_kwargs[k]["id"]
|
protocol_kwargs[k]["id"] if v == "unilabos_msgs/Resource" else protocol_kwargs[k][0]["id"]
|
||||||
if v == "unilabos_msgs/Resource"
|
|
||||||
else protocol_kwargs[k][0]["id"]
|
|
||||||
)
|
)
|
||||||
resource_uuid = protocol_kwargs[k].get("uuid", None)
|
resource_uuid = protocol_kwargs[k].get("uuid", None)
|
||||||
r = SerialCommand_Request()
|
r = SerialCommand_Request()
|
||||||
r.command = json.dumps({"id": resource_id, "uuid": resource_uuid, "with_children": True})
|
r.command = json.dumps({"id": resource_id, "uuid": resource_uuid, "with_children": True})
|
||||||
# 发送请求并等待响应
|
# 发送请求并等待响应
|
||||||
response: SerialCommand_Response = await self._resource_clients["resource_get"].call_async(
|
response: SerialCommand_Response = await self._resource_clients[
|
||||||
|
"resource_get"
|
||||||
|
].call_async(
|
||||||
r
|
r
|
||||||
) # type: ignore
|
) # type: ignore
|
||||||
raw_data = json.loads(response.response)
|
raw_data = json.loads(response.response)
|
||||||
@@ -304,52 +307,12 @@ class ROS2WorkstationNode(BaseROS2DeviceNode):
|
|||||||
|
|
||||||
# 向Host更新物料当前状态
|
# 向Host更新物料当前状态
|
||||||
for k, v in goal.get_fields_and_field_types().items():
|
for k, v in goal.get_fields_and_field_types().items():
|
||||||
if v not in ["unilabos_msgs/Resource", "sequence<unilabos_msgs/Resource>"]:
|
if v in ["unilabos_msgs/Resource", "sequence<unilabos_msgs/Resource>"]:
|
||||||
continue
|
r = ResourceUpdate.Request()
|
||||||
self.lab_logger().info(f"更新资源状态: {k}")
|
r.resources = [
|
||||||
try:
|
convert_to_ros_msg(Resource, rs) for rs in nested_dict_to_list(protocol_kwargs[k])
|
||||||
# 去重:使用 seen 集合获取唯一的资源对象
|
]
|
||||||
seen = set()
|
response = await self._resource_clients["resource_update"].call_async(r)
|
||||||
unique_resources = []
|
|
||||||
|
|
||||||
# 获取资源数据,统一转换为列表
|
|
||||||
resource_data = protocol_kwargs[k]
|
|
||||||
is_sequence = v != "unilabos_msgs/Resource"
|
|
||||||
if not is_sequence:
|
|
||||||
resource_list = [resource_data] if isinstance(resource_data, dict) else resource_data
|
|
||||||
else:
|
|
||||||
# 处理序列类型,可能是嵌套列表
|
|
||||||
resource_list = []
|
|
||||||
if isinstance(resource_data, list):
|
|
||||||
for item in resource_data:
|
|
||||||
if isinstance(item, list):
|
|
||||||
resource_list.extend(item)
|
|
||||||
else:
|
|
||||||
resource_list.append(item)
|
|
||||||
else:
|
|
||||||
resource_list = [resource_data]
|
|
||||||
|
|
||||||
for res_data in resource_list:
|
|
||||||
if not isinstance(res_data, dict):
|
|
||||||
continue
|
|
||||||
res_name = res_data.get("id") or res_data.get("name")
|
|
||||||
if not res_name:
|
|
||||||
continue
|
|
||||||
|
|
||||||
# 使用 resource_tracker 获取本地 PLR 实例
|
|
||||||
plr = self.resource_tracker.figure_resource({"name": res_name}, try_mode=False)
|
|
||||||
# 获取父资源
|
|
||||||
res = self.resource_tracker.parent_resource(plr)
|
|
||||||
if id(res) not in seen:
|
|
||||||
seen.add(id(res))
|
|
||||||
unique_resources.append(res)
|
|
||||||
|
|
||||||
# 使用新的资源树接口更新
|
|
||||||
if unique_resources:
|
|
||||||
await self.update_resource(unique_resources)
|
|
||||||
except Exception as e:
|
|
||||||
self.lab_logger().error(f"资源更新失败: {e}")
|
|
||||||
self.lab_logger().error(traceback.format_exc())
|
|
||||||
|
|
||||||
# 设置成功状态和返回值
|
# 设置成功状态和返回值
|
||||||
execution_success = True
|
execution_success = True
|
||||||
|
|||||||
182
unilabos/ros/x/rclpyx.py
Normal file
182
unilabos/ros/x/rclpyx.py
Normal file
@@ -0,0 +1,182 @@
|
|||||||
|
import asyncio
|
||||||
|
from asyncio import events
|
||||||
|
import threading
|
||||||
|
|
||||||
|
import rclpy
|
||||||
|
from rclpy.impl.implementation_singleton import rclpy_implementation as _rclpy
|
||||||
|
from rclpy.executors import await_or_execute, Executor
|
||||||
|
from rclpy.action import ActionClient, ActionServer
|
||||||
|
from rclpy.action.server import ServerGoalHandle, GoalResponse, GoalInfo, GoalStatus
|
||||||
|
from std_msgs.msg import String
|
||||||
|
from action_tutorials_interfaces.action import Fibonacci
|
||||||
|
|
||||||
|
|
||||||
|
loop = None
|
||||||
|
|
||||||
|
def get_event_loop():
|
||||||
|
global loop
|
||||||
|
return loop
|
||||||
|
|
||||||
|
|
||||||
|
async def default_handle_accepted_callback_async(goal_handle):
|
||||||
|
"""Execute the goal."""
|
||||||
|
await goal_handle.execute()
|
||||||
|
|
||||||
|
|
||||||
|
class ServerGoalHandleX(ServerGoalHandle):
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
async def execute(self, execute_callback=None):
|
||||||
|
# It's possible that there has been a request to cancel the goal prior to executing.
|
||||||
|
# In this case we want to avoid the illegal state transition to EXECUTING
|
||||||
|
# but still call the users execute callback to let them handle canceling the goal.
|
||||||
|
if not self.is_cancel_requested:
|
||||||
|
self._update_state(_rclpy.GoalEvent.EXECUTE)
|
||||||
|
await self._action_server.notify_execute_async(self, execute_callback)
|
||||||
|
|
||||||
|
|
||||||
|
class ActionServerX(ActionServer):
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
self.register_handle_accepted_callback(default_handle_accepted_callback_async)
|
||||||
|
|
||||||
|
async def _execute_goal_request(self, request_header_and_message):
|
||||||
|
request_header, goal_request = request_header_and_message
|
||||||
|
goal_uuid = goal_request.goal_id
|
||||||
|
goal_info = GoalInfo()
|
||||||
|
goal_info.goal_id = goal_uuid
|
||||||
|
|
||||||
|
self._node.get_logger().debug('New goal request with ID: {0}'.format(goal_uuid.uuid))
|
||||||
|
|
||||||
|
# Check if goal ID is already being tracked by this action server
|
||||||
|
with self._lock:
|
||||||
|
goal_id_exists = self._handle.goal_exists(goal_info)
|
||||||
|
|
||||||
|
accepted = False
|
||||||
|
if not goal_id_exists:
|
||||||
|
# Call user goal callback
|
||||||
|
response = await await_or_execute(self._goal_callback, goal_request.goal)
|
||||||
|
if not isinstance(response, GoalResponse):
|
||||||
|
self._node.get_logger().warning(
|
||||||
|
'Goal request callback did not return a GoalResponse type. Rejecting goal.')
|
||||||
|
else:
|
||||||
|
accepted = GoalResponse.ACCEPT == response
|
||||||
|
|
||||||
|
if accepted:
|
||||||
|
# Stamp time of acceptance
|
||||||
|
goal_info.stamp = self._node.get_clock().now().to_msg()
|
||||||
|
|
||||||
|
# Create a goal handle
|
||||||
|
try:
|
||||||
|
with self._lock:
|
||||||
|
goal_handle = ServerGoalHandleX(self, goal_info, goal_request.goal)
|
||||||
|
except RuntimeError as e:
|
||||||
|
self._node.get_logger().error(
|
||||||
|
'Failed to accept new goal with ID {0}: {1}'.format(goal_uuid.uuid, e))
|
||||||
|
accepted = False
|
||||||
|
else:
|
||||||
|
self._goal_handles[bytes(goal_uuid.uuid)] = goal_handle
|
||||||
|
|
||||||
|
# Send response
|
||||||
|
response_msg = self._action_type.Impl.SendGoalService.Response()
|
||||||
|
response_msg.accepted = accepted
|
||||||
|
response_msg.stamp = goal_info.stamp
|
||||||
|
self._handle.send_goal_response(request_header, response_msg)
|
||||||
|
|
||||||
|
if not accepted:
|
||||||
|
self._node.get_logger().debug('New goal rejected: {0}'.format(goal_uuid.uuid))
|
||||||
|
return
|
||||||
|
|
||||||
|
self._node.get_logger().debug('New goal accepted: {0}'.format(goal_uuid.uuid))
|
||||||
|
|
||||||
|
# Provide the user a reference to the goal handle
|
||||||
|
# await await_or_execute(self._handle_accepted_callback, goal_handle)
|
||||||
|
asyncio.create_task(self._handle_accepted_callback(goal_handle))
|
||||||
|
|
||||||
|
async def notify_execute_async(self, goal_handle, execute_callback):
|
||||||
|
# Use provided callback, defaulting to a previously registered callback
|
||||||
|
if execute_callback is None:
|
||||||
|
if self._execute_callback is None:
|
||||||
|
return
|
||||||
|
execute_callback = self._execute_callback
|
||||||
|
|
||||||
|
# Schedule user callback for execution
|
||||||
|
self._node.get_logger().info(f"{events.get_running_loop()}")
|
||||||
|
asyncio.create_task(self._execute_goal(execute_callback, goal_handle))
|
||||||
|
# loop = asyncio.new_event_loop()
|
||||||
|
# asyncio.set_event_loop(loop)
|
||||||
|
# task = loop.create_task(self._execute_goal(execute_callback, goal_handle))
|
||||||
|
# await task
|
||||||
|
|
||||||
|
|
||||||
|
class ActionClientX(ActionClient):
|
||||||
|
feedback_queue = asyncio.Queue()
|
||||||
|
|
||||||
|
async def feedback_cb(self, msg):
|
||||||
|
await self.feedback_queue.put(msg)
|
||||||
|
|
||||||
|
async def send_goal_async(self, goal_msg):
|
||||||
|
goal_future = super().send_goal_async(
|
||||||
|
goal_msg,
|
||||||
|
feedback_callback=self.feedback_cb
|
||||||
|
)
|
||||||
|
client_goal_handle = await asyncio.ensure_future(goal_future)
|
||||||
|
if not client_goal_handle.accepted:
|
||||||
|
raise Exception("Goal rejected.")
|
||||||
|
result_future = client_goal_handle.get_result_async()
|
||||||
|
while True:
|
||||||
|
feedback_future = asyncio.ensure_future(self.feedback_queue.get())
|
||||||
|
tasks = [result_future, feedback_future]
|
||||||
|
await asyncio.wait(tasks, return_when=asyncio.FIRST_COMPLETED)
|
||||||
|
if result_future.done():
|
||||||
|
result = result_future.result().result
|
||||||
|
yield (None, result)
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
feedback = feedback_future.result().feedback
|
||||||
|
yield (feedback, None)
|
||||||
|
|
||||||
|
|
||||||
|
async def main(node):
|
||||||
|
print('Node started.')
|
||||||
|
action_client = ActionClientX(node, Fibonacci, 'fibonacci')
|
||||||
|
goal_msg = Fibonacci.Goal()
|
||||||
|
goal_msg.order = 10
|
||||||
|
async for (feedback, result) in action_client.send_goal_async(goal_msg):
|
||||||
|
if feedback:
|
||||||
|
print(f'Feedback: {feedback}')
|
||||||
|
else:
|
||||||
|
print(f'Result: {result}')
|
||||||
|
print('Finished.')
|
||||||
|
|
||||||
|
|
||||||
|
async def ros_loop_node(node):
|
||||||
|
while rclpy.ok():
|
||||||
|
rclpy.spin_once(node, timeout_sec=0)
|
||||||
|
await asyncio.sleep(1e-4)
|
||||||
|
|
||||||
|
|
||||||
|
async def ros_loop(executor: Executor):
|
||||||
|
while rclpy.ok():
|
||||||
|
executor.spin_once(timeout_sec=0)
|
||||||
|
await asyncio.sleep(1e-4)
|
||||||
|
|
||||||
|
|
||||||
|
def run_event_loop():
|
||||||
|
global loop
|
||||||
|
loop = asyncio.new_event_loop()
|
||||||
|
asyncio.set_event_loop(loop)
|
||||||
|
loop.run_forever()
|
||||||
|
|
||||||
|
|
||||||
|
def run_event_loop_in_thread():
|
||||||
|
thread = threading.Thread(target=run_event_loop, args=())
|
||||||
|
thread.start()
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
rclpy.init()
|
||||||
|
node = rclpy.create_node('async_subscriber')
|
||||||
|
future = asyncio.wait([ros_loop(node), main()])
|
||||||
|
asyncio.get_event_loop().run_until_complete(future)
|
||||||
187
unilabos/utils/README_LOGGING.md
Normal file
187
unilabos/utils/README_LOGGING.md
Normal file
@@ -0,0 +1,187 @@
|
|||||||
|
# UniLabOS 日志配置说明
|
||||||
|
|
||||||
|
> **文件位置**: `unilabos/utils/log.py`
|
||||||
|
> **最后更新**: 2026-01-11
|
||||||
|
> **维护者**: Uni-Lab-OS 开发团队
|
||||||
|
|
||||||
|
本文档说明 UniLabOS 日志系统中对第三方库和内部模块的日志级别配置,避免控制台被过多的 DEBUG 日志淹没。
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 📋 已屏蔽的日志
|
||||||
|
|
||||||
|
以下库/模块的日志已被设置为 **WARNING** 或 **INFO** 级别,不再显示 DEBUG 日志:
|
||||||
|
|
||||||
|
### 1. pymodbus(Modbus 通信库)
|
||||||
|
|
||||||
|
**配置位置**: `log.py` 第196-200行
|
||||||
|
|
||||||
|
```python
|
||||||
|
# pymodbus 库的日志太详细,设置为 WARNING
|
||||||
|
logging.getLogger('pymodbus').setLevel(logging.WARNING)
|
||||||
|
logging.getLogger('pymodbus.logging').setLevel(logging.WARNING)
|
||||||
|
logging.getLogger('pymodbus.logging.base').setLevel(logging.WARNING)
|
||||||
|
logging.getLogger('pymodbus.logging.decoders').setLevel(logging.WARNING)
|
||||||
|
```
|
||||||
|
|
||||||
|
**屏蔽原因**:
|
||||||
|
- pymodbus 在 DEBUG 级别会输出每一次 Modbus 通信的详细信息
|
||||||
|
- 包括 `Processing: 0x5 0x1e 0x0 0x0...` 等原始数据
|
||||||
|
- 包括 `decoded PDU function_code(3 sub -1) -> ReadHoldingRegistersResponse(...)` 等解码信息
|
||||||
|
- 这些信息对日常使用价值不大,但会快速刷屏
|
||||||
|
|
||||||
|
**典型被屏蔽的日志**:
|
||||||
|
```
|
||||||
|
[DEBUG] Processing: 0x5 0x1e 0x0 0x0 0x0 0x7 0x1 0x3 0x4 0x0 0x0 0x0 0x0 [handleFrame:72] [pymodbus.logging.base]
|
||||||
|
[DEBUG] decoded PDU function_code(3 sub -1) -> ReadHoldingRegistersResponse(...) [decode:79] [pymodbus.logging.decoders]
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
### 2. websockets(WebSocket 库)
|
||||||
|
|
||||||
|
**配置位置**: `log.py` 第202-205行
|
||||||
|
|
||||||
|
```python
|
||||||
|
# websockets 库的日志输出较多,设置为 WARNING
|
||||||
|
logging.getLogger('websockets').setLevel(logging.WARNING)
|
||||||
|
logging.getLogger('websockets.client').setLevel(logging.WARNING)
|
||||||
|
logging.getLogger('websockets.server').setLevel(logging.WARNING)
|
||||||
|
```
|
||||||
|
|
||||||
|
**屏蔽原因**:
|
||||||
|
- WebSocket 连接、断开、心跳等信息在 DEBUG 级别会频繁输出
|
||||||
|
- 对于长时间运行的服务,这些日志意义不大
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
### 3. ROS Host Node(设备状态更新)
|
||||||
|
|
||||||
|
**配置位置**: `log.py` 第207-208行
|
||||||
|
|
||||||
|
```python
|
||||||
|
# ROS 节点的状态更新日志过于频繁,设置为 INFO
|
||||||
|
logging.getLogger('unilabos.ros.nodes.presets.host_node').setLevel(logging.INFO)
|
||||||
|
```
|
||||||
|
|
||||||
|
**屏蔽原因**:
|
||||||
|
- 设备状态更新(如手套箱压力)每隔几秒就会更新一次
|
||||||
|
- DEBUG 日志会记录每一次状态变化,导致日志刷屏
|
||||||
|
- 这些频繁的状态更新对调试价值不大
|
||||||
|
|
||||||
|
**典型被屏蔽的日志**:
|
||||||
|
```
|
||||||
|
[DEBUG] [/devices/host_node] Status updated: BatteryStation.data_glove_box_pressure = 4.229457855224609 [property_callback:666] [unilabos.ros.nodes.presets.host_node]
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
### 4. asyncio 和 urllib3
|
||||||
|
|
||||||
|
**配置位置**: `log.py` 第224-225行
|
||||||
|
|
||||||
|
```python
|
||||||
|
logging.getLogger("asyncio").setLevel(logging.INFO)
|
||||||
|
logging.getLogger("urllib3").setLevel(logging.INFO)
|
||||||
|
```
|
||||||
|
|
||||||
|
**屏蔽原因**:
|
||||||
|
- asyncio: 异步 IO 的内部调试信息
|
||||||
|
- urllib3: HTTP 请求库的连接池、重试等详细信息
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 🔧 如何临时启用这些日志(调试用)
|
||||||
|
|
||||||
|
### 方法1: 修改 log.py(永久启用)
|
||||||
|
|
||||||
|
在 `log.py` 的 `configure_logger()` 函数中,将对应库的日志级别改为 `logging.DEBUG`:
|
||||||
|
|
||||||
|
```python
|
||||||
|
# 临时启用 pymodbus 的 DEBUG 日志
|
||||||
|
logging.getLogger('pymodbus').setLevel(logging.DEBUG)
|
||||||
|
logging.getLogger('pymodbus.logging').setLevel(logging.DEBUG)
|
||||||
|
logging.getLogger('pymodbus.logging.base').setLevel(logging.DEBUG)
|
||||||
|
logging.getLogger('pymodbus.logging.decoders').setLevel(logging.DEBUG)
|
||||||
|
```
|
||||||
|
|
||||||
|
### 方法2: 在代码中临时启用(单次调试)
|
||||||
|
|
||||||
|
在需要调试的代码文件中添加:
|
||||||
|
|
||||||
|
```python
|
||||||
|
import logging
|
||||||
|
|
||||||
|
# 临时启用 pymodbus DEBUG 日志
|
||||||
|
logging.getLogger('pymodbus').setLevel(logging.DEBUG)
|
||||||
|
|
||||||
|
# 你的 Modbus 调试代码
|
||||||
|
...
|
||||||
|
|
||||||
|
# 调试完成后恢复
|
||||||
|
logging.getLogger('pymodbus').setLevel(logging.WARNING)
|
||||||
|
```
|
||||||
|
|
||||||
|
### 方法3: 使用环境变量或配置文件(推荐)
|
||||||
|
|
||||||
|
未来可以考虑在启动参数中添加 `--debug-modbus` 等选项来动态控制。
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 📊 日志级别说明
|
||||||
|
|
||||||
|
| 级别 | 数值 | 用途 | 是否显示 |
|
||||||
|
|------|------|------|---------|
|
||||||
|
| TRACE | 5 | 最详细的跟踪信息 | ✅ |
|
||||||
|
| DEBUG | 10 | 调试信息 | ✅ |
|
||||||
|
| INFO | 20 | 一般信息 | ✅ |
|
||||||
|
| WARNING | 30 | 警告信息 | ✅ |
|
||||||
|
| ERROR | 40 | 错误信息 | ✅ |
|
||||||
|
| CRITICAL | 50 | 严重错误 | ✅ |
|
||||||
|
|
||||||
|
**当前配置**:
|
||||||
|
- UniLabOS 自身代码: DEBUG 及以上全部显示
|
||||||
|
- pymodbus/websockets: **WARNING** 及以上显示(屏蔽 DEBUG/INFO)
|
||||||
|
- ROS host_node: **INFO** 及以上显示(屏蔽 DEBUG)
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## ⚠️ 重要提示
|
||||||
|
|
||||||
|
### 修改生效时间
|
||||||
|
- 修改 `log.py` 后需要 **重启 unilab 服务** 才能生效
|
||||||
|
- 不需要重新安装或重新编译
|
||||||
|
|
||||||
|
### 调试 Modbus 通信问题
|
||||||
|
如果需要调试 Modbus 通信故障,应该:
|
||||||
|
1. 临时启用 pymodbus DEBUG 日志(方法2)
|
||||||
|
2. 复现问题
|
||||||
|
3. 查看详细的通信日志
|
||||||
|
4. 调试完成后记得恢复 WARNING 级别
|
||||||
|
|
||||||
|
### 调试设备状态问题
|
||||||
|
如果需要调试设备状态更新问题:
|
||||||
|
```python
|
||||||
|
logging.getLogger('unilabos.ros.nodes.presets.host_node').setLevel(logging.DEBUG)
|
||||||
|
```
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 📝 维护记录
|
||||||
|
|
||||||
|
| 日期 | 修改内容 | 操作人 |
|
||||||
|
|------|---------|--------|
|
||||||
|
| 2026-01-11 | 初始创建,添加 pymodbus、websockets、ROS host_node 屏蔽 | - |
|
||||||
|
| 2026-01-07 | 添加 pymodbus 和 websockets 屏蔽(log-0107.py) | - |
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## 🔗 相关文件
|
||||||
|
|
||||||
|
- `log.py` - 日志配置主文件
|
||||||
|
- `unilabos/devices/workstation/coin_cell_assembly/` - 使用 Modbus 的扣电工作站代码
|
||||||
|
- `unilabos/ros/nodes/presets/host_node.py` - ROS 主机节点代码
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
**维护提示**: 如果添加了新的第三方库或发现新的日志刷屏问题,请在此文档中记录并更新 `log.py` 配置。
|
||||||
@@ -24,7 +24,6 @@ class EnvironmentChecker:
|
|||||||
"msgcenterpy": "msgcenterpy",
|
"msgcenterpy": "msgcenterpy",
|
||||||
"opentrons_shared_data": "opentrons_shared_data",
|
"opentrons_shared_data": "opentrons_shared_data",
|
||||||
"typing_extensions": "typing_extensions",
|
"typing_extensions": "typing_extensions",
|
||||||
"crcmod": "crcmod-plus",
|
|
||||||
}
|
}
|
||||||
|
|
||||||
# 特殊安装包(需要特殊处理的包)
|
# 特殊安装包(需要特殊处理的包)
|
||||||
|
|||||||
@@ -1,11 +1,7 @@
|
|||||||
import psutil
|
import psutil
|
||||||
import pywinauto
|
import pywinauto
|
||||||
try:
|
from pywinauto_recorder import UIApplication
|
||||||
from pywinauto_recorder import UIApplication
|
from pywinauto_recorder.player import UIPath, click, focus_on_application, exists, find, get_wrapper_path
|
||||||
from pywinauto_recorder.player import UIPath, click, focus_on_application, exists, find, get_wrapper_path
|
|
||||||
except ImportError:
|
|
||||||
print("未安装pywinauto_recorder,部分功能无法使用,安装时注意enum")
|
|
||||||
pass
|
|
||||||
from pywinauto.controls.uiawrapper import UIAWrapper
|
from pywinauto.controls.uiawrapper import UIAWrapper
|
||||||
from pywinauto.application import WindowSpecification
|
from pywinauto.application import WindowSpecification
|
||||||
from pywinauto import findbestmatch
|
from pywinauto import findbestmatch
|
||||||
|
|||||||
@@ -1,18 +0,0 @@
|
|||||||
networkx
|
|
||||||
typing_extensions
|
|
||||||
websockets
|
|
||||||
msgcenterpy>=0.1.5
|
|
||||||
opentrons_shared_data
|
|
||||||
pint
|
|
||||||
fastapi
|
|
||||||
jinja2
|
|
||||||
requests
|
|
||||||
uvicorn
|
|
||||||
pyautogui
|
|
||||||
opcua
|
|
||||||
pyserial
|
|
||||||
pandas
|
|
||||||
crcmod-plus
|
|
||||||
pymodbus
|
|
||||||
matplotlib
|
|
||||||
pylibftdi
|
|
||||||
@@ -1,89 +1,3 @@
|
|||||||
"""
|
|
||||||
工作流转换模块 - JSON 到 WorkflowGraph 的转换流程
|
|
||||||
|
|
||||||
==================== 输入格式 (JSON) ====================
|
|
||||||
|
|
||||||
{
|
|
||||||
"workflow": [
|
|
||||||
{"action": "transfer_liquid", "action_args": {"sources": "cell_lines", "targets": "Liquid_1", "asp_vol": 100.0, "dis_vol": 74.75, ...}},
|
|
||||||
...
|
|
||||||
],
|
|
||||||
"reagent": {
|
|
||||||
"cell_lines": {"slot": 4, "well": ["A1", "A3", "A5"], "labware": "DRUG + YOYO-MEDIA"},
|
|
||||||
"Liquid_1": {"slot": 1, "well": ["A4", "A7", "A10"], "labware": "rep 1"},
|
|
||||||
...
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
==================== 转换步骤 ====================
|
|
||||||
|
|
||||||
第一步: 按 slot 去重创建 create_resource 节点(创建板子)
|
|
||||||
--------------------------------------------------------------------------------
|
|
||||||
- 遍历所有 reagent,按 slot 去重,为每个唯一的 slot 创建一个板子
|
|
||||||
- 生成参数:
|
|
||||||
res_id: plate_slot_{slot}
|
|
||||||
device_id: /PRCXI
|
|
||||||
class_name: PRCXI_BioER_96_wellplate
|
|
||||||
parent: /PRCXI/PRCXI_Deck/T{slot}
|
|
||||||
slot_on_deck: "{slot}"
|
|
||||||
- 输出端口: labware(用于连接 set_liquid_from_plate)
|
|
||||||
- 控制流: create_resource 之间通过 ready 端口串联
|
|
||||||
|
|
||||||
示例: slot=1, slot=4 -> 创建 2 个 create_resource 节点
|
|
||||||
|
|
||||||
第二步: 为每个 reagent 创建 set_liquid_from_plate 节点(设置液体)
|
|
||||||
--------------------------------------------------------------------------------
|
|
||||||
- 遍历所有 reagent,为每个试剂创建 set_liquid_from_plate 节点
|
|
||||||
- 生成参数:
|
|
||||||
plate: [](通过连接传递,来自 create_resource 的 labware)
|
|
||||||
well_names: ["A1", "A3", "A5"](来自 reagent 的 well 数组)
|
|
||||||
liquid_names: ["cell_lines", "cell_lines", "cell_lines"](与 well 数量一致)
|
|
||||||
volumes: [1e5, 1e5, 1e5](与 well 数量一致,默认体积)
|
|
||||||
- 输入连接: create_resource (labware) -> set_liquid_from_plate (input_plate)
|
|
||||||
- 输出端口: output_wells(用于连接 transfer_liquid)
|
|
||||||
- 控制流: set_liquid_from_plate 连接在所有 create_resource 之后,通过 ready 端口串联
|
|
||||||
|
|
||||||
第三步: 解析 workflow,创建 transfer_liquid 等动作节点
|
|
||||||
--------------------------------------------------------------------------------
|
|
||||||
- 遍历 workflow 数组,为每个动作创建步骤节点
|
|
||||||
- 参数重命名: asp_vol -> asp_vols, dis_vol -> dis_vols, asp_flow_rate -> asp_flow_rates, dis_flow_rate -> dis_flow_rates
|
|
||||||
- 参数扩展: 根据 targets 的 wells 数量,将单值扩展为数组
|
|
||||||
例: asp_vol=100.0, targets 有 3 个 wells -> asp_vols=[100.0, 100.0, 100.0]
|
|
||||||
- 连接处理: 如果 sources/targets 已通过 set_liquid_from_plate 连接,参数值改为 []
|
|
||||||
- 输入连接: set_liquid_from_plate (output_wells) -> transfer_liquid (sources_identifier / targets_identifier)
|
|
||||||
- 输出端口: sources_out, targets_out(用于连接下一个 transfer_liquid)
|
|
||||||
|
|
||||||
==================== 连接关系图 ====================
|
|
||||||
|
|
||||||
控制流 (ready 端口串联):
|
|
||||||
create_resource_1 -> create_resource_2 -> ... -> set_liquid_1 -> set_liquid_2 -> ... -> transfer_liquid_1 -> transfer_liquid_2 -> ...
|
|
||||||
|
|
||||||
物料流:
|
|
||||||
[create_resource] --labware--> [set_liquid_from_plate] --output_wells--> [transfer_liquid] --sources_out/targets_out--> [下一个 transfer_liquid]
|
|
||||||
(slot=1) (cell_lines) (input_plate) (sources_identifier) (sources_identifier)
|
|
||||||
(slot=4) (Liquid_1) (targets_identifier) (targets_identifier)
|
|
||||||
|
|
||||||
==================== 端口映射 ====================
|
|
||||||
|
|
||||||
create_resource:
|
|
||||||
输出: labware
|
|
||||||
|
|
||||||
set_liquid_from_plate:
|
|
||||||
输入: input_plate
|
|
||||||
输出: output_plate, output_wells
|
|
||||||
|
|
||||||
transfer_liquid:
|
|
||||||
输入: sources -> sources_identifier, targets -> targets_identifier
|
|
||||||
输出: sources -> sources_out, targets -> targets_out
|
|
||||||
|
|
||||||
==================== 校验规则 ====================
|
|
||||||
|
|
||||||
- 检查 sources/targets 是否在 reagent 中定义
|
|
||||||
- 检查 sources 和 targets 的 wells 数量是否匹配
|
|
||||||
- 检查参数数组长度是否与 wells 数量一致
|
|
||||||
- 如有问题,在 footer 中添加 [WARN: ...] 标记
|
|
||||||
"""
|
|
||||||
|
|
||||||
import re
|
import re
|
||||||
import uuid
|
import uuid
|
||||||
|
|
||||||
@@ -94,28 +8,6 @@ from typing import Dict, List, Any, Tuple, Optional
|
|||||||
|
|
||||||
Json = Dict[str, Any]
|
Json = Dict[str, Any]
|
||||||
|
|
||||||
|
|
||||||
# ==================== 默认配置 ====================
|
|
||||||
|
|
||||||
# create_resource 节点默认参数
|
|
||||||
CREATE_RESOURCE_DEFAULTS = {
|
|
||||||
"device_id": "/PRCXI",
|
|
||||||
"parent_template": "/PRCXI/PRCXI_Deck/T{slot}", # {slot} 会被替换为实际的 slot 值
|
|
||||||
"class_name": "PRCXI_BioER_96_wellplate",
|
|
||||||
}
|
|
||||||
|
|
||||||
# 默认液体体积 (uL)
|
|
||||||
DEFAULT_LIQUID_VOLUME = 1e5
|
|
||||||
|
|
||||||
# 参数重命名映射:单数 -> 复数(用于 transfer_liquid 等动作)
|
|
||||||
PARAM_RENAME_MAPPING = {
|
|
||||||
"asp_vol": "asp_vols",
|
|
||||||
"dis_vol": "dis_vols",
|
|
||||||
"asp_flow_rate": "asp_flow_rates",
|
|
||||||
"dis_flow_rate": "dis_flow_rates",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
# ---------------- Graph ----------------
|
# ---------------- Graph ----------------
|
||||||
|
|
||||||
|
|
||||||
@@ -336,7 +228,7 @@ def refactor_data(
|
|||||||
|
|
||||||
|
|
||||||
def build_protocol_graph(
|
def build_protocol_graph(
|
||||||
labware_info: Dict[str, Dict[str, Any]],
|
labware_info: List[Dict[str, Any]],
|
||||||
protocol_steps: List[Dict[str, Any]],
|
protocol_steps: List[Dict[str, Any]],
|
||||||
workstation_name: str,
|
workstation_name: str,
|
||||||
action_resource_mapping: Optional[Dict[str, str]] = None,
|
action_resource_mapping: Optional[Dict[str, str]] = None,
|
||||||
@@ -344,227 +236,112 @@ def build_protocol_graph(
|
|||||||
"""统一的协议图构建函数,根据设备类型自动选择构建逻辑
|
"""统一的协议图构建函数,根据设备类型自动选择构建逻辑
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
labware_info: labware 信息字典,格式为 {name: {slot, well, labware, ...}, ...}
|
labware_info: labware 信息字典
|
||||||
protocol_steps: 协议步骤列表
|
protocol_steps: 协议步骤列表
|
||||||
workstation_name: 工作站名称
|
workstation_name: 工作站名称
|
||||||
action_resource_mapping: action 到 resource_name 的映射字典,可选
|
action_resource_mapping: action 到 resource_name 的映射字典,可选
|
||||||
"""
|
"""
|
||||||
G = WorkflowGraph()
|
G = WorkflowGraph()
|
||||||
resource_last_writer = {} # reagent_name -> "node_id:port"
|
resource_last_writer = {}
|
||||||
slot_to_create_resource = {} # slot -> create_resource node_id
|
|
||||||
|
|
||||||
protocol_steps = refactor_data(protocol_steps, action_resource_mapping)
|
protocol_steps = refactor_data(protocol_steps, action_resource_mapping)
|
||||||
|
# 有机化学&移液站协议图构建
|
||||||
|
WORKSTATION_ID = workstation_name
|
||||||
|
|
||||||
# ==================== 第一步:按 slot 去重创建 create_resource 节点 ====================
|
# 为所有labware创建资源节点
|
||||||
# 收集所有唯一的 slot
|
|
||||||
slots_info = {} # slot -> {labware, res_id}
|
|
||||||
for labware_id, item in labware_info.items():
|
|
||||||
slot = str(item.get("slot", ""))
|
|
||||||
if slot and slot not in slots_info:
|
|
||||||
res_id = f"plate_slot_{slot}"
|
|
||||||
slots_info[slot] = {
|
|
||||||
"labware": item.get("labware", ""),
|
|
||||||
"res_id": res_id,
|
|
||||||
}
|
|
||||||
|
|
||||||
# 为每个唯一的 slot 创建 create_resource 节点
|
|
||||||
res_index = 0
|
res_index = 0
|
||||||
last_create_resource_id = None
|
for labware_id, item in labware_info.items():
|
||||||
for slot, info in slots_info.items():
|
# item_id = item.get("id") or item.get("name", f"item_{uuid.uuid4()}")
|
||||||
node_id = str(uuid.uuid4())
|
node_id = str(uuid.uuid4())
|
||||||
res_id = info["res_id"]
|
|
||||||
|
# 判断节点类型
|
||||||
|
if "Rack" in str(labware_id) or "Tip" in str(labware_id):
|
||||||
|
lab_node_type = "Labware"
|
||||||
|
description = f"Prepare Labware: {labware_id}"
|
||||||
|
liquid_type = []
|
||||||
|
liquid_volume = []
|
||||||
|
elif item.get("type") == "hardware" or "reactor" in str(labware_id).lower():
|
||||||
|
if "reactor" not in str(labware_id).lower():
|
||||||
|
continue
|
||||||
|
lab_node_type = "Sample"
|
||||||
|
description = f"Prepare Reactor: {labware_id}"
|
||||||
|
liquid_type = []
|
||||||
|
liquid_volume = []
|
||||||
|
else:
|
||||||
|
lab_node_type = "Reagent"
|
||||||
|
description = f"Add Reagent to Flask: {labware_id}"
|
||||||
|
liquid_type = [labware_id]
|
||||||
|
liquid_volume = [1e5]
|
||||||
|
|
||||||
res_index += 1
|
res_index += 1
|
||||||
G.add_node(
|
G.add_node(
|
||||||
node_id,
|
node_id,
|
||||||
template_name="create_resource",
|
template_name="create_resource",
|
||||||
resource_name="host_node",
|
resource_name="host_node",
|
||||||
name=f"Plate {res_index}",
|
name=f"Res {res_index}",
|
||||||
description=f"Create plate on slot {slot}",
|
description=description,
|
||||||
lab_node_type="Labware",
|
lab_node_type=lab_node_type,
|
||||||
footer="create_resource-host_node",
|
footer="create_resource-host_node",
|
||||||
param={
|
param={
|
||||||
"res_id": res_id,
|
"res_id": labware_id,
|
||||||
"device_id": CREATE_RESOURCE_DEFAULTS["device_id"],
|
"device_id": WORKSTATION_ID,
|
||||||
"class_name": CREATE_RESOURCE_DEFAULTS["class_name"],
|
"class_name": "container",
|
||||||
"parent": CREATE_RESOURCE_DEFAULTS["parent_template"].format(slot=slot),
|
"parent": WORKSTATION_ID,
|
||||||
"bind_locations": {"x": 0.0, "y": 0.0, "z": 0.0},
|
"bind_locations": {"x": 0.0, "y": 0.0, "z": 0.0},
|
||||||
"slot_on_deck": slot,
|
"liquid_input_slot": [-1],
|
||||||
|
"liquid_type": liquid_type,
|
||||||
|
"liquid_volume": liquid_volume,
|
||||||
|
"slot_on_deck": "",
|
||||||
},
|
},
|
||||||
)
|
)
|
||||||
slot_to_create_resource[slot] = node_id
|
resource_last_writer[labware_id] = f"{node_id}:labware"
|
||||||
|
|
||||||
# create_resource 之间通过 ready 串联
|
last_control_node_id = None
|
||||||
if last_create_resource_id is not None:
|
|
||||||
G.add_edge(last_create_resource_id, node_id, source_port="ready", target_port="ready")
|
|
||||||
last_create_resource_id = node_id
|
|
||||||
|
|
||||||
# ==================== 第二步:为每个 reagent 创建 set_liquid_from_plate 节点 ====================
|
|
||||||
set_liquid_index = 0
|
|
||||||
last_set_liquid_id = last_create_resource_id # set_liquid_from_plate 连接在 create_resource 之后
|
|
||||||
|
|
||||||
for labware_id, item in labware_info.items():
|
|
||||||
# 跳过 Tip/Rack 类型
|
|
||||||
if "Rack" in str(labware_id) or "Tip" in str(labware_id):
|
|
||||||
continue
|
|
||||||
if item.get("type") == "hardware":
|
|
||||||
continue
|
|
||||||
|
|
||||||
slot = str(item.get("slot", ""))
|
|
||||||
wells = item.get("well", [])
|
|
||||||
if not wells or not slot:
|
|
||||||
continue
|
|
||||||
|
|
||||||
# res_id 不能有空格
|
|
||||||
res_id = str(labware_id).replace(" ", "_")
|
|
||||||
well_count = len(wells)
|
|
||||||
|
|
||||||
node_id = str(uuid.uuid4())
|
|
||||||
set_liquid_index += 1
|
|
||||||
|
|
||||||
G.add_node(
|
|
||||||
node_id,
|
|
||||||
template_name="set_liquid_from_plate",
|
|
||||||
resource_name="liquid_handler.prcxi",
|
|
||||||
name=f"SetLiquid {set_liquid_index}",
|
|
||||||
description=f"Set liquid: {labware_id}",
|
|
||||||
lab_node_type="Reagent",
|
|
||||||
footer="set_liquid_from_plate-liquid_handler.prcxi",
|
|
||||||
param={
|
|
||||||
"plate": [], # 通过连接传递
|
|
||||||
"well_names": wells, # 孔位名数组,如 ["A1", "A3", "A5"]
|
|
||||||
"liquid_names": [res_id] * well_count,
|
|
||||||
"volumes": [DEFAULT_LIQUID_VOLUME] * well_count,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
|
|
||||||
# ready 连接:上一个节点 -> set_liquid_from_plate
|
|
||||||
if last_set_liquid_id is not None:
|
|
||||||
G.add_edge(last_set_liquid_id, node_id, source_port="ready", target_port="ready")
|
|
||||||
last_set_liquid_id = node_id
|
|
||||||
|
|
||||||
# 物料流:create_resource 的 labware -> set_liquid_from_plate 的 input_plate
|
|
||||||
create_res_node_id = slot_to_create_resource.get(slot)
|
|
||||||
if create_res_node_id:
|
|
||||||
G.add_edge(create_res_node_id, node_id, source_port="labware", target_port="input_plate")
|
|
||||||
|
|
||||||
# set_liquid_from_plate 的输出 output_wells 用于连接 transfer_liquid
|
|
||||||
resource_last_writer[labware_id] = f"{node_id}:output_wells"
|
|
||||||
|
|
||||||
last_control_node_id = last_set_liquid_id
|
|
||||||
|
|
||||||
# 端口名称映射:JSON 字段名 -> 实际 handle key
|
|
||||||
INPUT_PORT_MAPPING = {
|
|
||||||
"sources": "sources_identifier",
|
|
||||||
"targets": "targets_identifier",
|
|
||||||
"vessel": "vessel",
|
|
||||||
"to_vessel": "to_vessel",
|
|
||||||
"from_vessel": "from_vessel",
|
|
||||||
"reagent": "reagent",
|
|
||||||
"solvent": "solvent",
|
|
||||||
"compound": "compound",
|
|
||||||
}
|
|
||||||
|
|
||||||
OUTPUT_PORT_MAPPING = {
|
|
||||||
"sources": "sources_out", # 输出端口是 xxx_out
|
|
||||||
"targets": "targets_out", # 输出端口是 xxx_out
|
|
||||||
"vessel": "vessel_out",
|
|
||||||
"to_vessel": "to_vessel_out",
|
|
||||||
"from_vessel": "from_vessel_out",
|
|
||||||
"filtrate_vessel": "filtrate_out",
|
|
||||||
"reagent": "reagent",
|
|
||||||
"solvent": "solvent",
|
|
||||||
"compound": "compound",
|
|
||||||
}
|
|
||||||
|
|
||||||
# 需要根据 wells 数量扩展的参数列表(复数形式)
|
|
||||||
EXPAND_BY_WELLS_PARAMS = ["asp_vols", "dis_vols", "asp_flow_rates", "dis_flow_rates"]
|
|
||||||
|
|
||||||
# 处理协议步骤
|
# 处理协议步骤
|
||||||
for step in protocol_steps:
|
for step in protocol_steps:
|
||||||
node_id = str(uuid.uuid4())
|
node_id = str(uuid.uuid4())
|
||||||
params = step.get("param", {}).copy() # 复制一份,避免修改原数据
|
G.add_node(node_id, **step)
|
||||||
connected_params = set() # 记录被连接的参数
|
|
||||||
warnings = [] # 收集警告信息
|
|
||||||
|
|
||||||
# 参数重命名:单数 -> 复数
|
|
||||||
for old_name, new_name in PARAM_RENAME_MAPPING.items():
|
|
||||||
if old_name in params:
|
|
||||||
params[new_name] = params.pop(old_name)
|
|
||||||
|
|
||||||
# 处理输入连接
|
|
||||||
for param_key, target_port in INPUT_PORT_MAPPING.items():
|
|
||||||
resource_name = params.get(param_key)
|
|
||||||
if resource_name and resource_name in resource_last_writer:
|
|
||||||
source_node, source_port = resource_last_writer[resource_name].split(":")
|
|
||||||
G.add_edge(source_node, node_id, source_port=source_port, target_port=target_port)
|
|
||||||
connected_params.add(param_key)
|
|
||||||
elif resource_name and resource_name not in resource_last_writer:
|
|
||||||
# 资源名在 labware_info 中不存在
|
|
||||||
warnings.append(f"{param_key}={resource_name} 未找到")
|
|
||||||
|
|
||||||
# 获取 targets 对应的 wells 数量,用于扩展参数
|
|
||||||
targets_name = params.get("targets")
|
|
||||||
sources_name = params.get("sources")
|
|
||||||
targets_wells_count = 1
|
|
||||||
sources_wells_count = 1
|
|
||||||
|
|
||||||
if targets_name and targets_name in labware_info:
|
|
||||||
target_wells = labware_info[targets_name].get("well", [])
|
|
||||||
targets_wells_count = len(target_wells) if target_wells else 1
|
|
||||||
elif targets_name:
|
|
||||||
warnings.append(f"targets={targets_name} 未在 reagent 中定义")
|
|
||||||
|
|
||||||
if sources_name and sources_name in labware_info:
|
|
||||||
source_wells = labware_info[sources_name].get("well", [])
|
|
||||||
sources_wells_count = len(source_wells) if source_wells else 1
|
|
||||||
elif sources_name:
|
|
||||||
warnings.append(f"sources={sources_name} 未在 reagent 中定义")
|
|
||||||
|
|
||||||
# 检查 sources 和 targets 的 wells 数量是否匹配
|
|
||||||
if targets_wells_count != sources_wells_count and targets_name and sources_name:
|
|
||||||
warnings.append(f"wells 数量不匹配: sources={sources_wells_count}, targets={targets_wells_count}")
|
|
||||||
|
|
||||||
# 使用 targets 的 wells 数量来扩展参数
|
|
||||||
wells_count = targets_wells_count
|
|
||||||
|
|
||||||
# 扩展单值参数为数组(根据 targets 的 wells 数量)
|
|
||||||
for expand_param in EXPAND_BY_WELLS_PARAMS:
|
|
||||||
if expand_param in params:
|
|
||||||
value = params[expand_param]
|
|
||||||
# 如果是单个值,扩展为数组
|
|
||||||
if not isinstance(value, list):
|
|
||||||
params[expand_param] = [value] * wells_count
|
|
||||||
# 如果已经是数组但长度不对,记录警告
|
|
||||||
elif len(value) != wells_count:
|
|
||||||
warnings.append(f"{expand_param} 数量({len(value)})与 wells({wells_count})不匹配")
|
|
||||||
|
|
||||||
# 如果 sources/targets 已通过连接传递,将参数值改为空数组
|
|
||||||
for param_key in connected_params:
|
|
||||||
if param_key in params:
|
|
||||||
params[param_key] = []
|
|
||||||
|
|
||||||
# 更新 step 的 param 和 footer
|
|
||||||
step_copy = step.copy()
|
|
||||||
step_copy["param"] = params
|
|
||||||
|
|
||||||
# 如果有警告,修改 footer 添加警告标记(警告放前面)
|
|
||||||
if warnings:
|
|
||||||
original_footer = step.get("footer", "")
|
|
||||||
step_copy["footer"] = f"[WARN: {'; '.join(warnings)}] {original_footer}"
|
|
||||||
|
|
||||||
G.add_node(node_id, **step_copy)
|
|
||||||
|
|
||||||
# 控制流
|
# 控制流
|
||||||
if last_control_node_id is not None:
|
if last_control_node_id is not None:
|
||||||
G.add_edge(last_control_node_id, node_id, source_port="ready", target_port="ready")
|
G.add_edge(last_control_node_id, node_id, source_port="ready", target_port="ready")
|
||||||
last_control_node_id = node_id
|
last_control_node_id = node_id
|
||||||
|
|
||||||
# 处理输出:更新 resource_last_writer
|
# 物料流
|
||||||
for param_key, output_port in OUTPUT_PORT_MAPPING.items():
|
params = step.get("param", {})
|
||||||
resource_name = step.get("param", {}).get(param_key) # 使用原始参数值
|
input_resources_possible_names = [
|
||||||
|
"vessel",
|
||||||
|
"to_vessel",
|
||||||
|
"from_vessel",
|
||||||
|
"reagent",
|
||||||
|
"solvent",
|
||||||
|
"compound",
|
||||||
|
"sources",
|
||||||
|
"targets",
|
||||||
|
]
|
||||||
|
|
||||||
|
for target_port in input_resources_possible_names:
|
||||||
|
resource_name = params.get(target_port)
|
||||||
|
if resource_name and resource_name in resource_last_writer:
|
||||||
|
source_node, source_port = resource_last_writer[resource_name].split(":")
|
||||||
|
G.add_edge(source_node, node_id, source_port=source_port, target_port=target_port)
|
||||||
|
|
||||||
|
output_resources = {
|
||||||
|
"vessel_out": params.get("vessel"),
|
||||||
|
"from_vessel_out": params.get("from_vessel"),
|
||||||
|
"to_vessel_out": params.get("to_vessel"),
|
||||||
|
"filtrate_out": params.get("filtrate_vessel"),
|
||||||
|
"reagent": params.get("reagent"),
|
||||||
|
"solvent": params.get("solvent"),
|
||||||
|
"compound": params.get("compound"),
|
||||||
|
"sources_out": params.get("sources"),
|
||||||
|
"targets_out": params.get("targets"),
|
||||||
|
}
|
||||||
|
|
||||||
|
for source_port, resource_name in output_resources.items():
|
||||||
if resource_name:
|
if resource_name:
|
||||||
resource_last_writer[resource_name] = f"{node_id}:{output_port}"
|
resource_last_writer[resource_name] = f"{node_id}:{source_port}"
|
||||||
|
|
||||||
return G
|
return G
|
||||||
|
|
||||||
|
|||||||
@@ -1,68 +1,21 @@
|
|||||||
"""
|
"""
|
||||||
JSON 工作流转换模块
|
JSON 工作流转换模块
|
||||||
|
|
||||||
将 workflow/reagent 格式的 JSON 转换为统一工作流格式。
|
提供从多种 JSON 格式转换为统一工作流格式的功能。
|
||||||
|
支持的格式:
|
||||||
输入格式:
|
1. workflow/reagent 格式
|
||||||
{
|
2. steps_info/labware_info 格式
|
||||||
"workflow": [
|
|
||||||
{"action": "...", "action_args": {...}},
|
|
||||||
...
|
|
||||||
],
|
|
||||||
"reagent": {
|
|
||||||
"reagent_name": {"slot": int, "well": [...], "labware": "..."},
|
|
||||||
...
|
|
||||||
}
|
|
||||||
}
|
|
||||||
"""
|
"""
|
||||||
|
|
||||||
import json
|
import json
|
||||||
from os import PathLike
|
from os import PathLike
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
from typing import Any, Dict, List, Optional, Tuple, Union
|
from typing import Any, Dict, List, Optional, Set, Tuple, Union
|
||||||
|
|
||||||
from unilabos.workflow.common import WorkflowGraph, build_protocol_graph
|
from unilabos.workflow.common import WorkflowGraph, build_protocol_graph
|
||||||
from unilabos.registry.registry import lab_registry
|
from unilabos.registry.registry import lab_registry
|
||||||
|
|
||||||
|
|
||||||
# ==================== 字段映射配置 ====================
|
|
||||||
|
|
||||||
# action 到 resource_name 的映射
|
|
||||||
ACTION_RESOURCE_MAPPING: Dict[str, str] = {
|
|
||||||
# 生物实验操作
|
|
||||||
"transfer_liquid": "liquid_handler.prcxi",
|
|
||||||
"transfer": "liquid_handler.prcxi",
|
|
||||||
"incubation": "incubator.prcxi",
|
|
||||||
"move_labware": "labware_mover.prcxi",
|
|
||||||
"oscillation": "shaker.prcxi",
|
|
||||||
# 有机化学操作
|
|
||||||
"HeatChillToTemp": "heatchill.chemputer",
|
|
||||||
"StopHeatChill": "heatchill.chemputer",
|
|
||||||
"StartHeatChill": "heatchill.chemputer",
|
|
||||||
"HeatChill": "heatchill.chemputer",
|
|
||||||
"Dissolve": "stirrer.chemputer",
|
|
||||||
"Transfer": "liquid_handler.chemputer",
|
|
||||||
"Evaporate": "rotavap.chemputer",
|
|
||||||
"Recrystallize": "reactor.chemputer",
|
|
||||||
"Filter": "filter.chemputer",
|
|
||||||
"Dry": "dryer.chemputer",
|
|
||||||
"Add": "liquid_handler.chemputer",
|
|
||||||
}
|
|
||||||
|
|
||||||
# action_args 字段到 parameters 字段的映射
|
|
||||||
# 格式: {"old_key": "new_key"}, 仅映射需要重命名的字段
|
|
||||||
ARGS_FIELD_MAPPING: Dict[str, str] = {
|
|
||||||
# 如果需要字段重命名,在这里配置
|
|
||||||
# "old_field_name": "new_field_name",
|
|
||||||
}
|
|
||||||
|
|
||||||
# 默认工作站名称
|
|
||||||
DEFAULT_WORKSTATION = "PRCXI"
|
|
||||||
|
|
||||||
|
|
||||||
# ==================== 核心转换函数 ====================
|
|
||||||
|
|
||||||
|
|
||||||
def get_action_handles(resource_name: str, template_name: str) -> Dict[str, List[str]]:
|
def get_action_handles(resource_name: str, template_name: str) -> Dict[str, List[str]]:
|
||||||
"""
|
"""
|
||||||
从 registry 获取指定设备和动作的 handles 配置
|
从 registry 获取指定设备和动作的 handles 配置
|
||||||
@@ -86,10 +39,12 @@ def get_action_handles(resource_name: str, template_name: str) -> Dict[str, List
|
|||||||
handles = action_config.get("handles", {})
|
handles = action_config.get("handles", {})
|
||||||
|
|
||||||
if isinstance(handles, dict):
|
if isinstance(handles, dict):
|
||||||
|
# 处理 input handles (作为 target)
|
||||||
for handle in handles.get("input", []):
|
for handle in handles.get("input", []):
|
||||||
handler_key = handle.get("handler_key", "")
|
handler_key = handle.get("handler_key", "")
|
||||||
if handler_key:
|
if handler_key:
|
||||||
result["source"].append(handler_key)
|
result["source"].append(handler_key)
|
||||||
|
# 处理 output handles (作为 source)
|
||||||
for handle in handles.get("output", []):
|
for handle in handles.get("output", []):
|
||||||
handler_key = handle.get("handler_key", "")
|
handler_key = handle.get("handler_key", "")
|
||||||
if handler_key:
|
if handler_key:
|
||||||
@@ -114,9 +69,12 @@ def validate_workflow_handles(graph: WorkflowGraph) -> Tuple[bool, List[str]]:
|
|||||||
for edge in graph.edges:
|
for edge in graph.edges:
|
||||||
left_uuid = edge.get("source")
|
left_uuid = edge.get("source")
|
||||||
right_uuid = edge.get("target")
|
right_uuid = edge.get("target")
|
||||||
|
# target_handle_key是target, right的输入节点(入节点)
|
||||||
|
# source_handle_key是source, left的输出节点(出节点)
|
||||||
right_source_conn_key = edge.get("target_handle_key", "")
|
right_source_conn_key = edge.get("target_handle_key", "")
|
||||||
left_target_conn_key = edge.get("source_handle_key", "")
|
left_target_conn_key = edge.get("source_handle_key", "")
|
||||||
|
|
||||||
|
# 获取源节点和目标节点信息
|
||||||
left_node = nodes.get(left_uuid, {})
|
left_node = nodes.get(left_uuid, {})
|
||||||
right_node = nodes.get(right_uuid, {})
|
right_node = nodes.get(right_uuid, {})
|
||||||
|
|
||||||
@@ -125,93 +83,164 @@ def validate_workflow_handles(graph: WorkflowGraph) -> Tuple[bool, List[str]]:
|
|||||||
right_res_name = right_node.get("resource_name", "")
|
right_res_name = right_node.get("resource_name", "")
|
||||||
right_template_name = right_node.get("template_name", "")
|
right_template_name = right_node.get("template_name", "")
|
||||||
|
|
||||||
|
# 获取源节点的 output handles
|
||||||
left_node_handles = get_action_handles(left_res_name, left_template_name)
|
left_node_handles = get_action_handles(left_res_name, left_template_name)
|
||||||
target_valid_keys = left_node_handles.get("target", [])
|
target_valid_keys = left_node_handles.get("target", [])
|
||||||
target_valid_keys.append("ready")
|
target_valid_keys.append("ready")
|
||||||
|
|
||||||
|
# 获取目标节点的 input handles
|
||||||
right_node_handles = get_action_handles(right_res_name, right_template_name)
|
right_node_handles = get_action_handles(right_res_name, right_template_name)
|
||||||
source_valid_keys = right_node_handles.get("source", [])
|
source_valid_keys = right_node_handles.get("source", [])
|
||||||
source_valid_keys.append("ready")
|
source_valid_keys.append("ready")
|
||||||
|
|
||||||
# 验证目标节点(right)的输入端口
|
# 如果节点配置了 output handles,则 source_port 必须有效
|
||||||
if not right_source_conn_key:
|
if not right_source_conn_key:
|
||||||
node_name = right_node.get("name", right_uuid[:8])
|
node_name = left_node.get("name", left_uuid[:8])
|
||||||
errors.append(f"目标节点 '{node_name}' 的输入端口 (target_handle_key) 为空,应设置为: {source_valid_keys}")
|
errors.append(f"源节点 '{node_name}' 的 source_handle_key 为空," f"应设置为: {source_valid_keys}")
|
||||||
elif right_source_conn_key not in source_valid_keys:
|
elif right_source_conn_key not in source_valid_keys:
|
||||||
node_name = right_node.get("name", right_uuid[:8])
|
node_name = left_node.get("name", left_uuid[:8])
|
||||||
errors.append(
|
errors.append(
|
||||||
f"目标节点 '{node_name}' 的输入端口 '{right_source_conn_key}' 不存在,支持的输入端口: {source_valid_keys}"
|
f"源节点 '{node_name}' 的 source 端点 '{right_source_conn_key}' 不存在," f"支持的端点: {source_valid_keys}"
|
||||||
)
|
)
|
||||||
|
|
||||||
# 验证源节点(left)的输出端口
|
# 如果节点配置了 input handles,则 target_port 必须有效
|
||||||
if not left_target_conn_key:
|
if not left_target_conn_key:
|
||||||
node_name = left_node.get("name", left_uuid[:8])
|
node_name = right_node.get("name", right_uuid[:8])
|
||||||
errors.append(f"源节点 '{node_name}' 的输出端口 (source_handle_key) 为空,应设置为: {target_valid_keys}")
|
errors.append(f"目标节点 '{node_name}' 的 target_handle_key 为空," f"应设置为: {target_valid_keys}")
|
||||||
elif left_target_conn_key not in target_valid_keys:
|
elif left_target_conn_key not in target_valid_keys:
|
||||||
node_name = left_node.get("name", left_uuid[:8])
|
node_name = right_node.get("name", right_uuid[:8])
|
||||||
errors.append(
|
errors.append(
|
||||||
f"源节点 '{node_name}' 的输出端口 '{left_target_conn_key}' 不存在,支持的输出端口: {target_valid_keys}"
|
f"目标节点 '{node_name}' 的 target 端点 '{left_target_conn_key}' 不存在,"
|
||||||
|
f"支持的端点: {target_valid_keys}"
|
||||||
)
|
)
|
||||||
|
|
||||||
return len(errors) == 0, errors
|
return len(errors) == 0, errors
|
||||||
|
|
||||||
|
|
||||||
def normalize_workflow_steps(workflow: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
# action 到 resource_name 的映射
|
||||||
|
ACTION_RESOURCE_MAPPING: Dict[str, str] = {
|
||||||
|
# 生物实验操作
|
||||||
|
"transfer_liquid": "liquid_handler.prcxi",
|
||||||
|
"transfer": "liquid_handler.prcxi",
|
||||||
|
"incubation": "incubator.prcxi",
|
||||||
|
"move_labware": "labware_mover.prcxi",
|
||||||
|
"oscillation": "shaker.prcxi",
|
||||||
|
# 有机化学操作
|
||||||
|
"HeatChillToTemp": "heatchill.chemputer",
|
||||||
|
"StopHeatChill": "heatchill.chemputer",
|
||||||
|
"StartHeatChill": "heatchill.chemputer",
|
||||||
|
"HeatChill": "heatchill.chemputer",
|
||||||
|
"Dissolve": "stirrer.chemputer",
|
||||||
|
"Transfer": "liquid_handler.chemputer",
|
||||||
|
"Evaporate": "rotavap.chemputer",
|
||||||
|
"Recrystallize": "reactor.chemputer",
|
||||||
|
"Filter": "filter.chemputer",
|
||||||
|
"Dry": "dryer.chemputer",
|
||||||
|
"Add": "liquid_handler.chemputer",
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def normalize_steps(data: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
||||||
"""
|
"""
|
||||||
将 workflow 格式的步骤数据规范化
|
将不同格式的步骤数据规范化为统一格式
|
||||||
|
|
||||||
输入格式:
|
支持的输入格式:
|
||||||
[{"action": "...", "action_args": {...}}, ...]
|
- action + parameters
|
||||||
|
- action + action_args
|
||||||
输出格式:
|
- operation + parameters
|
||||||
[{"action": "...", "parameters": {...}, "step_number": int}, ...]
|
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
workflow: workflow 数组
|
data: 原始步骤数据列表
|
||||||
|
|
||||||
Returns:
|
Returns:
|
||||||
规范化后的步骤列表
|
规范化后的步骤列表,格式为 [{"action": str, "parameters": dict, "description": str?, "step_number": int?}, ...]
|
||||||
"""
|
"""
|
||||||
normalized = []
|
normalized = []
|
||||||
for idx, step in enumerate(workflow):
|
for idx, step in enumerate(data):
|
||||||
action = step.get("action")
|
# 获取动作名称(支持 action 或 operation 字段)
|
||||||
|
action = step.get("action") or step.get("operation")
|
||||||
if not action:
|
if not action:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# 获取参数: action_args
|
# 获取参数(支持 parameters 或 action_args 字段)
|
||||||
raw_params = step.get("action_args", {})
|
raw_params = step.get("parameters") or step.get("action_args") or {}
|
||||||
params = {}
|
params = dict(raw_params)
|
||||||
|
|
||||||
# 应用字段映射
|
# 规范化 source/target -> sources/targets
|
||||||
for key, value in raw_params.items():
|
if "source" in raw_params and "sources" not in raw_params:
|
||||||
mapped_key = ARGS_FIELD_MAPPING.get(key, key)
|
params["sources"] = raw_params["source"]
|
||||||
params[mapped_key] = value
|
if "target" in raw_params and "targets" not in raw_params:
|
||||||
|
params["targets"] = raw_params["target"]
|
||||||
|
|
||||||
step_dict = {
|
# 获取描述(支持 description 或 purpose 字段)
|
||||||
"action": action,
|
description = step.get("description") or step.get("purpose")
|
||||||
"parameters": params,
|
|
||||||
"step_number": idx + 1,
|
|
||||||
}
|
|
||||||
|
|
||||||
# 保留描述字段
|
# 获取步骤编号(优先使用原始数据中的 step_number,否则使用索引+1)
|
||||||
if "description" in step:
|
step_number = step.get("step_number", idx + 1)
|
||||||
step_dict["description"] = step["description"]
|
|
||||||
|
step_dict = {"action": action, "parameters": params, "step_number": step_number}
|
||||||
|
if description:
|
||||||
|
step_dict["description"] = description
|
||||||
|
|
||||||
normalized.append(step_dict)
|
normalized.append(step_dict)
|
||||||
|
|
||||||
return normalized
|
return normalized
|
||||||
|
|
||||||
|
|
||||||
|
def normalize_labware(data: List[Dict[str, Any]]) -> Dict[str, Dict[str, Any]]:
|
||||||
|
"""
|
||||||
|
将不同格式的 labware 数据规范化为统一的字典格式
|
||||||
|
|
||||||
|
支持的输入格式:
|
||||||
|
- reagent_name + material_name + positions
|
||||||
|
- name + labware + slot
|
||||||
|
|
||||||
|
Args:
|
||||||
|
data: 原始 labware 数据列表
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
规范化后的 labware 字典,格式为 {name: {"slot": int, "labware": str, "well": list, "type": str, "role": str, "name": str}, ...}
|
||||||
|
"""
|
||||||
|
labware = {}
|
||||||
|
for item in data:
|
||||||
|
# 获取 key 名称(优先使用 reagent_name,其次是 material_name 或 name)
|
||||||
|
reagent_name = item.get("reagent_name")
|
||||||
|
key = reagent_name or item.get("material_name") or item.get("name")
|
||||||
|
if not key:
|
||||||
|
continue
|
||||||
|
|
||||||
|
key = str(key)
|
||||||
|
|
||||||
|
# 处理重复 key,自动添加后缀
|
||||||
|
idx = 1
|
||||||
|
original_key = key
|
||||||
|
while key in labware:
|
||||||
|
idx += 1
|
||||||
|
key = f"{original_key}_{idx}"
|
||||||
|
|
||||||
|
labware[key] = {
|
||||||
|
"slot": item.get("positions") or item.get("slot"),
|
||||||
|
"labware": item.get("material_name") or item.get("labware"),
|
||||||
|
"well": item.get("well", []),
|
||||||
|
"type": item.get("type", "reagent"),
|
||||||
|
"role": item.get("role", ""),
|
||||||
|
"name": key,
|
||||||
|
}
|
||||||
|
|
||||||
|
return labware
|
||||||
|
|
||||||
|
|
||||||
def convert_from_json(
|
def convert_from_json(
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
data: Union[str, PathLike, Dict[str, Any]],
|
||||||
workstation_name: str = DEFAULT_WORKSTATION,
|
workstation_name: str = "PRCXi",
|
||||||
validate: bool = True,
|
validate: bool = True,
|
||||||
) -> WorkflowGraph:
|
) -> WorkflowGraph:
|
||||||
"""
|
"""
|
||||||
从 JSON 数据或文件转换为 WorkflowGraph
|
从 JSON 数据或文件转换为 WorkflowGraph
|
||||||
|
|
||||||
JSON 格式:
|
支持的 JSON 格式:
|
||||||
{"workflow": [...], "reagent": {...}}
|
1. {"workflow": [...], "reagent": {...}} - 直接格式
|
||||||
|
2. {"steps_info": [...], "labware_info": [...]} - 需要规范化的格式
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
data: JSON 文件路径、字典数据、或 JSON 字符串
|
data: JSON 文件路径、字典数据、或 JSON 字符串
|
||||||
@@ -222,7 +251,7 @@ def convert_from_json(
|
|||||||
WorkflowGraph: 构建好的工作流图
|
WorkflowGraph: 构建好的工作流图
|
||||||
|
|
||||||
Raises:
|
Raises:
|
||||||
ValueError: 不支持的 JSON 格式
|
ValueError: 不支持的 JSON 格式 或 句柄校验失败
|
||||||
FileNotFoundError: 文件不存在
|
FileNotFoundError: 文件不存在
|
||||||
json.JSONDecodeError: JSON 解析失败
|
json.JSONDecodeError: JSON 解析失败
|
||||||
"""
|
"""
|
||||||
@@ -233,6 +262,7 @@ def convert_from_json(
|
|||||||
with path.open("r", encoding="utf-8") as fp:
|
with path.open("r", encoding="utf-8") as fp:
|
||||||
json_data = json.load(fp)
|
json_data = json.load(fp)
|
||||||
elif isinstance(data, str):
|
elif isinstance(data, str):
|
||||||
|
# 尝试作为 JSON 字符串解析
|
||||||
json_data = json.loads(data)
|
json_data = json.loads(data)
|
||||||
else:
|
else:
|
||||||
raise FileNotFoundError(f"文件不存在: {data}")
|
raise FileNotFoundError(f"文件不存在: {data}")
|
||||||
@@ -241,24 +271,30 @@ def convert_from_json(
|
|||||||
else:
|
else:
|
||||||
raise TypeError(f"不支持的数据类型: {type(data)}")
|
raise TypeError(f"不支持的数据类型: {type(data)}")
|
||||||
|
|
||||||
# 校验格式
|
# 根据格式解析数据
|
||||||
if "workflow" not in json_data or "reagent" not in json_data:
|
if "workflow" in json_data and "reagent" in json_data:
|
||||||
|
# 格式1: workflow/reagent(已经是规范格式)
|
||||||
|
protocol_steps = json_data["workflow"]
|
||||||
|
labware_info = json_data["reagent"]
|
||||||
|
elif "steps_info" in json_data and "labware_info" in json_data:
|
||||||
|
# 格式2: steps_info/labware_info(需要规范化)
|
||||||
|
protocol_steps = normalize_steps(json_data["steps_info"])
|
||||||
|
labware_info = normalize_labware(json_data["labware_info"])
|
||||||
|
elif "steps" in json_data and "labware" in json_data:
|
||||||
|
# 格式3: steps/labware(另一种常见格式)
|
||||||
|
protocol_steps = normalize_steps(json_data["steps"])
|
||||||
|
if isinstance(json_data["labware"], list):
|
||||||
|
labware_info = normalize_labware(json_data["labware"])
|
||||||
|
else:
|
||||||
|
labware_info = json_data["labware"]
|
||||||
|
else:
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"不支持的 JSON 格式。请使用标准格式:\n"
|
"不支持的 JSON 格式。支持的格式:\n"
|
||||||
'{"workflow": [{"action": "...", "action_args": {...}}, ...], '
|
"1. {'workflow': [...], 'reagent': {...}}\n"
|
||||||
'"reagent": {"name": {"slot": int, "well": [...], "labware": "..."}, ...}}'
|
"2. {'steps_info': [...], 'labware_info': [...]}\n"
|
||||||
|
"3. {'steps': [...], 'labware': [...]}"
|
||||||
)
|
)
|
||||||
|
|
||||||
# 提取数据
|
|
||||||
workflow = json_data["workflow"]
|
|
||||||
reagent = json_data["reagent"]
|
|
||||||
|
|
||||||
# 规范化步骤数据
|
|
||||||
protocol_steps = normalize_workflow_steps(workflow)
|
|
||||||
|
|
||||||
# reagent 已经是字典格式,直接使用
|
|
||||||
labware_info = reagent
|
|
||||||
|
|
||||||
# 构建工作流图
|
# 构建工作流图
|
||||||
graph = build_protocol_graph(
|
graph = build_protocol_graph(
|
||||||
labware_info=labware_info,
|
labware_info=labware_info,
|
||||||
@@ -281,7 +317,7 @@ def convert_from_json(
|
|||||||
|
|
||||||
def convert_json_to_node_link(
|
def convert_json_to_node_link(
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
data: Union[str, PathLike, Dict[str, Any]],
|
||||||
workstation_name: str = DEFAULT_WORKSTATION,
|
workstation_name: str = "PRCXi",
|
||||||
) -> Dict[str, Any]:
|
) -> Dict[str, Any]:
|
||||||
"""
|
"""
|
||||||
将 JSON 数据转换为 node-link 格式的字典
|
将 JSON 数据转换为 node-link 格式的字典
|
||||||
@@ -299,7 +335,7 @@ def convert_json_to_node_link(
|
|||||||
|
|
||||||
def convert_json_to_workflow_list(
|
def convert_json_to_workflow_list(
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
data: Union[str, PathLike, Dict[str, Any]],
|
||||||
workstation_name: str = DEFAULT_WORKSTATION,
|
workstation_name: str = "PRCXi",
|
||||||
) -> List[Dict[str, Any]]:
|
) -> List[Dict[str, Any]]:
|
||||||
"""
|
"""
|
||||||
将 JSON 数据转换为工作流列表格式
|
将 JSON 数据转换为工作流列表格式
|
||||||
@@ -313,3 +349,8 @@ def convert_json_to_workflow_list(
|
|||||||
"""
|
"""
|
||||||
graph = convert_from_json(data, workstation_name)
|
graph = convert_from_json(data, workstation_name)
|
||||||
return graph.to_dict()
|
return graph.to_dict()
|
||||||
|
|
||||||
|
|
||||||
|
# 为了向后兼容,保留下划线前缀的别名
|
||||||
|
_normalize_steps = normalize_steps
|
||||||
|
_normalize_labware = normalize_labware
|
||||||
|
|||||||
241
unilabos/workflow/from_python_script.py
Normal file
241
unilabos/workflow/from_python_script.py
Normal file
@@ -0,0 +1,241 @@
|
|||||||
|
import ast
|
||||||
|
import json
|
||||||
|
from typing import Dict, List, Any, Tuple, Optional
|
||||||
|
|
||||||
|
from .common import WorkflowGraph, RegistryAdapter
|
||||||
|
|
||||||
|
Json = Dict[str, Any]
|
||||||
|
|
||||||
|
# ---------------- Converter ----------------
|
||||||
|
|
||||||
|
class DeviceMethodConverter:
|
||||||
|
"""
|
||||||
|
- 字段统一:resource_name(原 device_class)、template_name(原 action_key)
|
||||||
|
- params 单层;inputs 使用 'params.' 前缀
|
||||||
|
- SimpleGraph.add_workflow_node 负责变量连线与边
|
||||||
|
"""
|
||||||
|
def __init__(self, device_registry: Optional[Dict[str, Any]] = None):
|
||||||
|
self.graph = WorkflowGraph()
|
||||||
|
self.variable_sources: Dict[str, Dict[str, Any]] = {} # var -> {node_id, output_name}
|
||||||
|
self.instance_to_resource: Dict[str, Optional[str]] = {} # 实例名 -> resource_name
|
||||||
|
self.node_id_counter: int = 0
|
||||||
|
self.registry = RegistryAdapter(device_registry or {})
|
||||||
|
|
||||||
|
# ---- helpers ----
|
||||||
|
def _new_node_id(self) -> int:
|
||||||
|
nid = self.node_id_counter
|
||||||
|
self.node_id_counter += 1
|
||||||
|
return nid
|
||||||
|
|
||||||
|
def _assign_targets(self, targets) -> List[str]:
|
||||||
|
names: List[str] = []
|
||||||
|
import ast
|
||||||
|
if isinstance(targets, ast.Tuple):
|
||||||
|
for elt in targets.elts:
|
||||||
|
if isinstance(elt, ast.Name):
|
||||||
|
names.append(elt.id)
|
||||||
|
elif isinstance(targets, ast.Name):
|
||||||
|
names.append(targets.id)
|
||||||
|
return names
|
||||||
|
|
||||||
|
def _extract_device_instantiation(self, node) -> Optional[Tuple[str, str]]:
|
||||||
|
import ast
|
||||||
|
if not isinstance(node.value, ast.Call):
|
||||||
|
return None
|
||||||
|
callee = node.value.func
|
||||||
|
if isinstance(callee, ast.Name):
|
||||||
|
class_name = callee.id
|
||||||
|
elif isinstance(callee, ast.Attribute) and isinstance(callee.value, ast.Name):
|
||||||
|
class_name = callee.attr
|
||||||
|
else:
|
||||||
|
return None
|
||||||
|
if isinstance(node.targets[0], ast.Name):
|
||||||
|
instance = node.targets[0].id
|
||||||
|
return instance, class_name
|
||||||
|
return None
|
||||||
|
|
||||||
|
def _extract_call(self, call) -> Tuple[str, str, Dict[str, Any], str]:
|
||||||
|
import ast
|
||||||
|
owner_name, method_name, call_kind = "", "", "func"
|
||||||
|
if isinstance(call.func, ast.Attribute):
|
||||||
|
method_name = call.func.attr
|
||||||
|
if isinstance(call.func.value, ast.Name):
|
||||||
|
owner_name = call.func.value.id
|
||||||
|
call_kind = "instance" if owner_name in self.instance_to_resource else "class_or_module"
|
||||||
|
elif isinstance(call.func.value, ast.Attribute) and isinstance(call.func.value.value, ast.Name):
|
||||||
|
owner_name = call.func.value.attr
|
||||||
|
call_kind = "class_or_module"
|
||||||
|
elif isinstance(call.func, ast.Name):
|
||||||
|
method_name = call.func.id
|
||||||
|
call_kind = "func"
|
||||||
|
|
||||||
|
def pack(node):
|
||||||
|
if isinstance(node, ast.Name):
|
||||||
|
return {"type": "variable", "value": node.id}
|
||||||
|
if isinstance(node, ast.Constant):
|
||||||
|
return {"type": "constant", "value": node.value}
|
||||||
|
if isinstance(node, ast.Dict):
|
||||||
|
return {"type": "dict", "value": self._parse_dict(node)}
|
||||||
|
if isinstance(node, ast.List):
|
||||||
|
return {"type": "list", "value": self._parse_list(node)}
|
||||||
|
return {"type": "raw", "value": ast.unparse(node) if hasattr(ast, "unparse") else str(node)}
|
||||||
|
|
||||||
|
args: Dict[str, Any] = {}
|
||||||
|
pos: List[Any] = []
|
||||||
|
for a in call.args:
|
||||||
|
pos.append(pack(a))
|
||||||
|
for kw in call.keywords:
|
||||||
|
args[kw.arg] = pack(kw.value)
|
||||||
|
if pos:
|
||||||
|
args["_positional"] = pos
|
||||||
|
return owner_name, method_name, args, call_kind
|
||||||
|
|
||||||
|
def _parse_dict(self, node) -> Dict[str, Any]:
|
||||||
|
import ast
|
||||||
|
out: Dict[str, Any] = {}
|
||||||
|
for k, v in zip(node.keys, node.values):
|
||||||
|
if isinstance(k, ast.Constant):
|
||||||
|
key = str(k.value)
|
||||||
|
if isinstance(v, ast.Name):
|
||||||
|
out[key] = f"var:{v.id}"
|
||||||
|
elif isinstance(v, ast.Constant):
|
||||||
|
out[key] = v.value
|
||||||
|
elif isinstance(v, ast.Dict):
|
||||||
|
out[key] = self._parse_dict(v)
|
||||||
|
elif isinstance(v, ast.List):
|
||||||
|
out[key] = self._parse_list(v)
|
||||||
|
return out
|
||||||
|
|
||||||
|
def _parse_list(self, node) -> List[Any]:
|
||||||
|
import ast
|
||||||
|
out: List[Any] = []
|
||||||
|
for elt in node.elts:
|
||||||
|
if isinstance(elt, ast.Name):
|
||||||
|
out.append(f"var:{elt.id}")
|
||||||
|
elif isinstance(elt, ast.Constant):
|
||||||
|
out.append(elt.value)
|
||||||
|
elif isinstance(elt, ast.Dict):
|
||||||
|
out.append(self._parse_dict(elt))
|
||||||
|
elif isinstance(elt, ast.List):
|
||||||
|
out.append(self._parse_list(elt))
|
||||||
|
return out
|
||||||
|
|
||||||
|
def _normalize_var_tokens(self, x: Any) -> Any:
|
||||||
|
if isinstance(x, str) and x.startswith("var:"):
|
||||||
|
return {"__var__": x[4:]}
|
||||||
|
if isinstance(x, list):
|
||||||
|
return [self._normalize_var_tokens(i) for i in x]
|
||||||
|
if isinstance(x, dict):
|
||||||
|
return {k: self._normalize_var_tokens(v) for k, v in x.items()}
|
||||||
|
return x
|
||||||
|
|
||||||
|
def _make_params_payload(self, resource_name: Optional[str], template_name: str, call_args: Dict[str, Any]) -> Dict[str, Any]:
|
||||||
|
input_keys = self.registry.get_action_input_keys(resource_name, template_name) if resource_name else []
|
||||||
|
defaults = self.registry.get_action_goal_default(resource_name, template_name) if resource_name else {}
|
||||||
|
params: Dict[str, Any] = dict(defaults)
|
||||||
|
|
||||||
|
def unpack(p):
|
||||||
|
t, v = p.get("type"), p.get("value")
|
||||||
|
if t == "variable":
|
||||||
|
return {"__var__": v}
|
||||||
|
if t == "dict":
|
||||||
|
return self._normalize_var_tokens(v)
|
||||||
|
if t == "list":
|
||||||
|
return self._normalize_var_tokens(v)
|
||||||
|
return v
|
||||||
|
|
||||||
|
for k, p in call_args.items():
|
||||||
|
if k == "_positional":
|
||||||
|
continue
|
||||||
|
params[k] = unpack(p)
|
||||||
|
|
||||||
|
pos = call_args.get("_positional", [])
|
||||||
|
if pos:
|
||||||
|
if input_keys:
|
||||||
|
for i, p in enumerate(pos):
|
||||||
|
if i >= len(input_keys):
|
||||||
|
break
|
||||||
|
name = input_keys[i]
|
||||||
|
if name in params:
|
||||||
|
continue
|
||||||
|
params[name] = unpack(p)
|
||||||
|
else:
|
||||||
|
for i, p in enumerate(pos):
|
||||||
|
params[f"arg_{i}"] = unpack(p)
|
||||||
|
return params
|
||||||
|
|
||||||
|
# ---- handlers ----
|
||||||
|
def _on_assign(self, stmt):
|
||||||
|
import ast
|
||||||
|
inst = self._extract_device_instantiation(stmt)
|
||||||
|
if inst:
|
||||||
|
instance, code_class = inst
|
||||||
|
resource_name = self.registry.resolve_resource_by_classname(code_class)
|
||||||
|
self.instance_to_resource[instance] = resource_name
|
||||||
|
return
|
||||||
|
|
||||||
|
if isinstance(stmt.value, ast.Call):
|
||||||
|
owner, method, call_args, kind = self._extract_call(stmt.value)
|
||||||
|
if kind == "instance":
|
||||||
|
device_key = owner
|
||||||
|
resource_name = self.instance_to_resource.get(owner)
|
||||||
|
else:
|
||||||
|
device_key = owner
|
||||||
|
resource_name = self.registry.resolve_resource_by_classname(owner)
|
||||||
|
|
||||||
|
module = self.registry.get_device_module(resource_name)
|
||||||
|
params = self._make_params_payload(resource_name, method, call_args)
|
||||||
|
|
||||||
|
nid = self._new_node_id()
|
||||||
|
self.graph.add_workflow_node(
|
||||||
|
nid,
|
||||||
|
device_key=device_key,
|
||||||
|
resource_name=resource_name, # ✅
|
||||||
|
module=module,
|
||||||
|
template_name=method, # ✅
|
||||||
|
params=params,
|
||||||
|
variable_sources=self.variable_sources,
|
||||||
|
add_ready_if_no_vars=True,
|
||||||
|
prev_node_id=(nid - 1) if nid > 0 else None,
|
||||||
|
)
|
||||||
|
|
||||||
|
out_vars = self._assign_targets(stmt.targets[0])
|
||||||
|
for var in out_vars:
|
||||||
|
self.variable_sources[var] = {"node_id": nid, "output_name": "result"}
|
||||||
|
|
||||||
|
def _on_expr(self, stmt):
|
||||||
|
import ast
|
||||||
|
if not isinstance(stmt.value, ast.Call):
|
||||||
|
return
|
||||||
|
owner, method, call_args, kind = self._extract_call(stmt.value)
|
||||||
|
if kind == "instance":
|
||||||
|
device_key = owner
|
||||||
|
resource_name = self.instance_to_resource.get(owner)
|
||||||
|
else:
|
||||||
|
device_key = owner
|
||||||
|
resource_name = self.registry.resolve_resource_by_classname(owner)
|
||||||
|
|
||||||
|
module = self.registry.get_device_module(resource_name)
|
||||||
|
params = self._make_params_payload(resource_name, method, call_args)
|
||||||
|
|
||||||
|
nid = self._new_node_id()
|
||||||
|
self.graph.add_workflow_node(
|
||||||
|
nid,
|
||||||
|
device_key=device_key,
|
||||||
|
resource_name=resource_name, # ✅
|
||||||
|
module=module,
|
||||||
|
template_name=method, # ✅
|
||||||
|
params=params,
|
||||||
|
variable_sources=self.variable_sources,
|
||||||
|
add_ready_if_no_vars=True,
|
||||||
|
prev_node_id=(nid - 1) if nid > 0 else None,
|
||||||
|
)
|
||||||
|
|
||||||
|
def convert(self, python_code: str):
|
||||||
|
tree = ast.parse(python_code)
|
||||||
|
for stmt in tree.body:
|
||||||
|
if isinstance(stmt, ast.Assign):
|
||||||
|
self._on_assign(stmt)
|
||||||
|
elif isinstance(stmt, ast.Expr):
|
||||||
|
self._on_expr(stmt)
|
||||||
|
return self
|
||||||
131
unilabos/workflow/from_xdl.py
Normal file
131
unilabos/workflow/from_xdl.py
Normal file
@@ -0,0 +1,131 @@
|
|||||||
|
from typing import List, Any, Dict
|
||||||
|
import xml.etree.ElementTree as ET
|
||||||
|
|
||||||
|
|
||||||
|
def convert_to_type(val: str) -> Any:
|
||||||
|
"""将字符串值转换为适当的数据类型"""
|
||||||
|
if val == "True":
|
||||||
|
return True
|
||||||
|
if val == "False":
|
||||||
|
return False
|
||||||
|
if val == "?":
|
||||||
|
return None
|
||||||
|
if val.endswith(" g"):
|
||||||
|
return float(val.split(" ")[0])
|
||||||
|
if val.endswith("mg"):
|
||||||
|
return float(val.split("mg")[0])
|
||||||
|
elif val.endswith("mmol"):
|
||||||
|
return float(val.split("mmol")[0]) / 1000
|
||||||
|
elif val.endswith("mol"):
|
||||||
|
return float(val.split("mol")[0])
|
||||||
|
elif val.endswith("ml"):
|
||||||
|
return float(val.split("ml")[0])
|
||||||
|
elif val.endswith("RPM"):
|
||||||
|
return float(val.split("RPM")[0])
|
||||||
|
elif val.endswith(" °C"):
|
||||||
|
return float(val.split(" ")[0])
|
||||||
|
elif val.endswith(" %"):
|
||||||
|
return float(val.split(" ")[0])
|
||||||
|
return val
|
||||||
|
|
||||||
|
|
||||||
|
def flatten_xdl_procedure(procedure_elem: ET.Element) -> List[ET.Element]:
|
||||||
|
"""展平嵌套的XDL程序结构"""
|
||||||
|
flattened_operations = []
|
||||||
|
TEMP_UNSUPPORTED_PROTOCOL = ["Purge", "Wait", "Stir", "ResetHandling"]
|
||||||
|
|
||||||
|
def extract_operations(element: ET.Element):
|
||||||
|
if element.tag not in ["Prep", "Reaction", "Workup", "Purification", "Procedure"]:
|
||||||
|
if element.tag not in TEMP_UNSUPPORTED_PROTOCOL:
|
||||||
|
flattened_operations.append(element)
|
||||||
|
|
||||||
|
for child in element:
|
||||||
|
extract_operations(child)
|
||||||
|
|
||||||
|
for child in procedure_elem:
|
||||||
|
extract_operations(child)
|
||||||
|
|
||||||
|
return flattened_operations
|
||||||
|
|
||||||
|
|
||||||
|
def parse_xdl_content(xdl_content: str) -> tuple:
|
||||||
|
"""解析XDL内容"""
|
||||||
|
try:
|
||||||
|
xdl_content_cleaned = "".join(c for c in xdl_content if c.isprintable())
|
||||||
|
root = ET.fromstring(xdl_content_cleaned)
|
||||||
|
|
||||||
|
synthesis_elem = root.find("Synthesis")
|
||||||
|
if synthesis_elem is None:
|
||||||
|
return None, None, None
|
||||||
|
|
||||||
|
# 解析硬件组件
|
||||||
|
hardware_elem = synthesis_elem.find("Hardware")
|
||||||
|
hardware = []
|
||||||
|
if hardware_elem is not None:
|
||||||
|
hardware = [{"id": c.get("id"), "type": c.get("type")} for c in hardware_elem.findall("Component")]
|
||||||
|
|
||||||
|
# 解析试剂
|
||||||
|
reagents_elem = synthesis_elem.find("Reagents")
|
||||||
|
reagents = []
|
||||||
|
if reagents_elem is not None:
|
||||||
|
reagents = [{"name": r.get("name"), "role": r.get("role", "")} for r in reagents_elem.findall("Reagent")]
|
||||||
|
|
||||||
|
# 解析程序
|
||||||
|
procedure_elem = synthesis_elem.find("Procedure")
|
||||||
|
if procedure_elem is None:
|
||||||
|
return None, None, None
|
||||||
|
|
||||||
|
flattened_operations = flatten_xdl_procedure(procedure_elem)
|
||||||
|
return hardware, reagents, flattened_operations
|
||||||
|
|
||||||
|
except ET.ParseError as e:
|
||||||
|
raise ValueError(f"Invalid XDL format: {e}")
|
||||||
|
|
||||||
|
|
||||||
|
def convert_xdl_to_dict(xdl_content: str) -> Dict[str, Any]:
|
||||||
|
"""
|
||||||
|
将XDL XML格式转换为标准的字典格式
|
||||||
|
|
||||||
|
Args:
|
||||||
|
xdl_content: XDL XML内容
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
转换结果,包含步骤和器材信息
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
hardware, reagents, flattened_operations = parse_xdl_content(xdl_content)
|
||||||
|
if hardware is None:
|
||||||
|
return {"error": "Failed to parse XDL content", "success": False}
|
||||||
|
|
||||||
|
# 将XDL元素转换为字典格式
|
||||||
|
steps_data = []
|
||||||
|
for elem in flattened_operations:
|
||||||
|
# 转换参数类型
|
||||||
|
parameters = {}
|
||||||
|
for key, val in elem.attrib.items():
|
||||||
|
converted_val = convert_to_type(val)
|
||||||
|
if converted_val is not None:
|
||||||
|
parameters[key] = converted_val
|
||||||
|
|
||||||
|
step_dict = {
|
||||||
|
"operation": elem.tag,
|
||||||
|
"parameters": parameters,
|
||||||
|
"description": elem.get("purpose", f"Operation: {elem.tag}"),
|
||||||
|
}
|
||||||
|
steps_data.append(step_dict)
|
||||||
|
|
||||||
|
# 合并硬件和试剂为统一的labware_info格式
|
||||||
|
labware_data = []
|
||||||
|
labware_data.extend({"id": hw["id"], "type": "hardware", **hw} for hw in hardware)
|
||||||
|
labware_data.extend({"name": reagent["name"], "type": "reagent", **reagent} for reagent in reagents)
|
||||||
|
|
||||||
|
return {
|
||||||
|
"success": True,
|
||||||
|
"steps": steps_data,
|
||||||
|
"labware": labware_data,
|
||||||
|
"message": f"Successfully converted XDL to dict format. Found {len(steps_data)} steps and {len(labware_data)} labware items.",
|
||||||
|
}
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
error_msg = f"XDL conversion failed: {str(e)}"
|
||||||
|
return {"error": error_msg, "success": False}
|
||||||
@@ -1,356 +0,0 @@
|
|||||||
"""
|
|
||||||
JSON 工作流转换模块
|
|
||||||
|
|
||||||
提供从多种 JSON 格式转换为统一工作流格式的功能。
|
|
||||||
支持的格式:
|
|
||||||
1. workflow/reagent 格式
|
|
||||||
2. steps_info/labware_info 格式
|
|
||||||
"""
|
|
||||||
|
|
||||||
import json
|
|
||||||
from os import PathLike
|
|
||||||
from pathlib import Path
|
|
||||||
from typing import Any, Dict, List, Optional, Set, Tuple, Union
|
|
||||||
|
|
||||||
from unilabos.workflow.common import WorkflowGraph, build_protocol_graph
|
|
||||||
from unilabos.registry.registry import lab_registry
|
|
||||||
|
|
||||||
|
|
||||||
def get_action_handles(resource_name: str, template_name: str) -> Dict[str, List[str]]:
|
|
||||||
"""
|
|
||||||
从 registry 获取指定设备和动作的 handles 配置
|
|
||||||
|
|
||||||
Args:
|
|
||||||
resource_name: 设备资源名称,如 "liquid_handler.prcxi"
|
|
||||||
template_name: 动作模板名称,如 "transfer_liquid"
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
包含 source 和 target handler_keys 的字典:
|
|
||||||
{"source": ["sources_out", "targets_out", ...], "target": ["sources", "targets", ...]}
|
|
||||||
"""
|
|
||||||
result = {"source": [], "target": []}
|
|
||||||
|
|
||||||
device_info = lab_registry.device_type_registry.get(resource_name, {})
|
|
||||||
if not device_info:
|
|
||||||
return result
|
|
||||||
|
|
||||||
action_mappings = device_info.get("class", {}).get("action_value_mappings", {})
|
|
||||||
action_config = action_mappings.get(template_name, {})
|
|
||||||
handles = action_config.get("handles", {})
|
|
||||||
|
|
||||||
if isinstance(handles, dict):
|
|
||||||
# 处理 input handles (作为 target)
|
|
||||||
for handle in handles.get("input", []):
|
|
||||||
handler_key = handle.get("handler_key", "")
|
|
||||||
if handler_key:
|
|
||||||
result["source"].append(handler_key)
|
|
||||||
# 处理 output handles (作为 source)
|
|
||||||
for handle in handles.get("output", []):
|
|
||||||
handler_key = handle.get("handler_key", "")
|
|
||||||
if handler_key:
|
|
||||||
result["target"].append(handler_key)
|
|
||||||
|
|
||||||
return result
|
|
||||||
|
|
||||||
|
|
||||||
def validate_workflow_handles(graph: WorkflowGraph) -> Tuple[bool, List[str]]:
|
|
||||||
"""
|
|
||||||
校验工作流图中所有边的句柄配置是否正确
|
|
||||||
|
|
||||||
Args:
|
|
||||||
graph: 工作流图对象
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
(is_valid, errors): 是否有效,错误信息列表
|
|
||||||
"""
|
|
||||||
errors = []
|
|
||||||
nodes = graph.nodes
|
|
||||||
|
|
||||||
for edge in graph.edges:
|
|
||||||
left_uuid = edge.get("source")
|
|
||||||
right_uuid = edge.get("target")
|
|
||||||
# target_handle_key是target, right的输入节点(入节点)
|
|
||||||
# source_handle_key是source, left的输出节点(出节点)
|
|
||||||
right_source_conn_key = edge.get("target_handle_key", "")
|
|
||||||
left_target_conn_key = edge.get("source_handle_key", "")
|
|
||||||
|
|
||||||
# 获取源节点和目标节点信息
|
|
||||||
left_node = nodes.get(left_uuid, {})
|
|
||||||
right_node = nodes.get(right_uuid, {})
|
|
||||||
|
|
||||||
left_res_name = left_node.get("resource_name", "")
|
|
||||||
left_template_name = left_node.get("template_name", "")
|
|
||||||
right_res_name = right_node.get("resource_name", "")
|
|
||||||
right_template_name = right_node.get("template_name", "")
|
|
||||||
|
|
||||||
# 获取源节点的 output handles
|
|
||||||
left_node_handles = get_action_handles(left_res_name, left_template_name)
|
|
||||||
target_valid_keys = left_node_handles.get("target", [])
|
|
||||||
target_valid_keys.append("ready")
|
|
||||||
|
|
||||||
# 获取目标节点的 input handles
|
|
||||||
right_node_handles = get_action_handles(right_res_name, right_template_name)
|
|
||||||
source_valid_keys = right_node_handles.get("source", [])
|
|
||||||
source_valid_keys.append("ready")
|
|
||||||
|
|
||||||
# 如果节点配置了 output handles,则 source_port 必须有效
|
|
||||||
if not right_source_conn_key:
|
|
||||||
node_name = left_node.get("name", left_uuid[:8])
|
|
||||||
errors.append(f"源节点 '{node_name}' 的 source_handle_key 为空," f"应设置为: {source_valid_keys}")
|
|
||||||
elif right_source_conn_key not in source_valid_keys:
|
|
||||||
node_name = left_node.get("name", left_uuid[:8])
|
|
||||||
errors.append(
|
|
||||||
f"源节点 '{node_name}' 的 source 端点 '{right_source_conn_key}' 不存在," f"支持的端点: {source_valid_keys}"
|
|
||||||
)
|
|
||||||
|
|
||||||
# 如果节点配置了 input handles,则 target_port 必须有效
|
|
||||||
if not left_target_conn_key:
|
|
||||||
node_name = right_node.get("name", right_uuid[:8])
|
|
||||||
errors.append(f"目标节点 '{node_name}' 的 target_handle_key 为空," f"应设置为: {target_valid_keys}")
|
|
||||||
elif left_target_conn_key not in target_valid_keys:
|
|
||||||
node_name = right_node.get("name", right_uuid[:8])
|
|
||||||
errors.append(
|
|
||||||
f"目标节点 '{node_name}' 的 target 端点 '{left_target_conn_key}' 不存在,"
|
|
||||||
f"支持的端点: {target_valid_keys}"
|
|
||||||
)
|
|
||||||
|
|
||||||
return len(errors) == 0, errors
|
|
||||||
|
|
||||||
|
|
||||||
# action 到 resource_name 的映射
|
|
||||||
ACTION_RESOURCE_MAPPING: Dict[str, str] = {
|
|
||||||
# 生物实验操作
|
|
||||||
"transfer_liquid": "liquid_handler.prcxi",
|
|
||||||
"transfer": "liquid_handler.prcxi",
|
|
||||||
"incubation": "incubator.prcxi",
|
|
||||||
"move_labware": "labware_mover.prcxi",
|
|
||||||
"oscillation": "shaker.prcxi",
|
|
||||||
# 有机化学操作
|
|
||||||
"HeatChillToTemp": "heatchill.chemputer",
|
|
||||||
"StopHeatChill": "heatchill.chemputer",
|
|
||||||
"StartHeatChill": "heatchill.chemputer",
|
|
||||||
"HeatChill": "heatchill.chemputer",
|
|
||||||
"Dissolve": "stirrer.chemputer",
|
|
||||||
"Transfer": "liquid_handler.chemputer",
|
|
||||||
"Evaporate": "rotavap.chemputer",
|
|
||||||
"Recrystallize": "reactor.chemputer",
|
|
||||||
"Filter": "filter.chemputer",
|
|
||||||
"Dry": "dryer.chemputer",
|
|
||||||
"Add": "liquid_handler.chemputer",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def normalize_steps(data: List[Dict[str, Any]]) -> List[Dict[str, Any]]:
|
|
||||||
"""
|
|
||||||
将不同格式的步骤数据规范化为统一格式
|
|
||||||
|
|
||||||
支持的输入格式:
|
|
||||||
- action + parameters
|
|
||||||
- action + action_args
|
|
||||||
- operation + parameters
|
|
||||||
|
|
||||||
Args:
|
|
||||||
data: 原始步骤数据列表
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
规范化后的步骤列表,格式为 [{"action": str, "parameters": dict, "description": str?, "step_number": int?}, ...]
|
|
||||||
"""
|
|
||||||
normalized = []
|
|
||||||
for idx, step in enumerate(data):
|
|
||||||
# 获取动作名称(支持 action 或 operation 字段)
|
|
||||||
action = step.get("action") or step.get("operation")
|
|
||||||
if not action:
|
|
||||||
continue
|
|
||||||
|
|
||||||
# 获取参数(支持 parameters 或 action_args 字段)
|
|
||||||
raw_params = step.get("parameters") or step.get("action_args") or {}
|
|
||||||
params = dict(raw_params)
|
|
||||||
|
|
||||||
# 规范化 source/target -> sources/targets
|
|
||||||
if "source" in raw_params and "sources" not in raw_params:
|
|
||||||
params["sources"] = raw_params["source"]
|
|
||||||
if "target" in raw_params and "targets" not in raw_params:
|
|
||||||
params["targets"] = raw_params["target"]
|
|
||||||
|
|
||||||
# 获取描述(支持 description 或 purpose 字段)
|
|
||||||
description = step.get("description") or step.get("purpose")
|
|
||||||
|
|
||||||
# 获取步骤编号(优先使用原始数据中的 step_number,否则使用索引+1)
|
|
||||||
step_number = step.get("step_number", idx + 1)
|
|
||||||
|
|
||||||
step_dict = {"action": action, "parameters": params, "step_number": step_number}
|
|
||||||
if description:
|
|
||||||
step_dict["description"] = description
|
|
||||||
|
|
||||||
normalized.append(step_dict)
|
|
||||||
|
|
||||||
return normalized
|
|
||||||
|
|
||||||
|
|
||||||
def normalize_labware(data: List[Dict[str, Any]]) -> Dict[str, Dict[str, Any]]:
|
|
||||||
"""
|
|
||||||
将不同格式的 labware 数据规范化为统一的字典格式
|
|
||||||
|
|
||||||
支持的输入格式:
|
|
||||||
- reagent_name + material_name + positions
|
|
||||||
- name + labware + slot
|
|
||||||
|
|
||||||
Args:
|
|
||||||
data: 原始 labware 数据列表
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
规范化后的 labware 字典,格式为 {name: {"slot": int, "labware": str, "well": list, "type": str, "role": str, "name": str}, ...}
|
|
||||||
"""
|
|
||||||
labware = {}
|
|
||||||
for item in data:
|
|
||||||
# 获取 key 名称(优先使用 reagent_name,其次是 material_name 或 name)
|
|
||||||
reagent_name = item.get("reagent_name")
|
|
||||||
key = reagent_name or item.get("material_name") or item.get("name")
|
|
||||||
if not key:
|
|
||||||
continue
|
|
||||||
|
|
||||||
key = str(key)
|
|
||||||
|
|
||||||
# 处理重复 key,自动添加后缀
|
|
||||||
idx = 1
|
|
||||||
original_key = key
|
|
||||||
while key in labware:
|
|
||||||
idx += 1
|
|
||||||
key = f"{original_key}_{idx}"
|
|
||||||
|
|
||||||
labware[key] = {
|
|
||||||
"slot": item.get("positions") or item.get("slot"),
|
|
||||||
"labware": item.get("material_name") or item.get("labware"),
|
|
||||||
"well": item.get("well", []),
|
|
||||||
"type": item.get("type", "reagent"),
|
|
||||||
"role": item.get("role", ""),
|
|
||||||
"name": key,
|
|
||||||
}
|
|
||||||
|
|
||||||
return labware
|
|
||||||
|
|
||||||
|
|
||||||
def convert_from_json(
|
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
|
||||||
workstation_name: str = "PRCXi",
|
|
||||||
validate: bool = True,
|
|
||||||
) -> WorkflowGraph:
|
|
||||||
"""
|
|
||||||
从 JSON 数据或文件转换为 WorkflowGraph
|
|
||||||
|
|
||||||
支持的 JSON 格式:
|
|
||||||
1. {"workflow": [...], "reagent": {...}} - 直接格式
|
|
||||||
2. {"steps_info": [...], "labware_info": [...]} - 需要规范化的格式
|
|
||||||
|
|
||||||
Args:
|
|
||||||
data: JSON 文件路径、字典数据、或 JSON 字符串
|
|
||||||
workstation_name: 工作站名称,默认 "PRCXi"
|
|
||||||
validate: 是否校验句柄配置,默认 True
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
WorkflowGraph: 构建好的工作流图
|
|
||||||
|
|
||||||
Raises:
|
|
||||||
ValueError: 不支持的 JSON 格式 或 句柄校验失败
|
|
||||||
FileNotFoundError: 文件不存在
|
|
||||||
json.JSONDecodeError: JSON 解析失败
|
|
||||||
"""
|
|
||||||
# 处理输入数据
|
|
||||||
if isinstance(data, (str, PathLike)):
|
|
||||||
path = Path(data)
|
|
||||||
if path.exists():
|
|
||||||
with path.open("r", encoding="utf-8") as fp:
|
|
||||||
json_data = json.load(fp)
|
|
||||||
elif isinstance(data, str):
|
|
||||||
# 尝试作为 JSON 字符串解析
|
|
||||||
json_data = json.loads(data)
|
|
||||||
else:
|
|
||||||
raise FileNotFoundError(f"文件不存在: {data}")
|
|
||||||
elif isinstance(data, dict):
|
|
||||||
json_data = data
|
|
||||||
else:
|
|
||||||
raise TypeError(f"不支持的数据类型: {type(data)}")
|
|
||||||
|
|
||||||
# 根据格式解析数据
|
|
||||||
if "workflow" in json_data and "reagent" in json_data:
|
|
||||||
# 格式1: workflow/reagent(已经是规范格式)
|
|
||||||
protocol_steps = json_data["workflow"]
|
|
||||||
labware_info = json_data["reagent"]
|
|
||||||
elif "steps_info" in json_data and "labware_info" in json_data:
|
|
||||||
# 格式2: steps_info/labware_info(需要规范化)
|
|
||||||
protocol_steps = normalize_steps(json_data["steps_info"])
|
|
||||||
labware_info = normalize_labware(json_data["labware_info"])
|
|
||||||
elif "steps" in json_data and "labware" in json_data:
|
|
||||||
# 格式3: steps/labware(另一种常见格式)
|
|
||||||
protocol_steps = normalize_steps(json_data["steps"])
|
|
||||||
if isinstance(json_data["labware"], list):
|
|
||||||
labware_info = normalize_labware(json_data["labware"])
|
|
||||||
else:
|
|
||||||
labware_info = json_data["labware"]
|
|
||||||
else:
|
|
||||||
raise ValueError(
|
|
||||||
"不支持的 JSON 格式。支持的格式:\n"
|
|
||||||
"1. {'workflow': [...], 'reagent': {...}}\n"
|
|
||||||
"2. {'steps_info': [...], 'labware_info': [...]}\n"
|
|
||||||
"3. {'steps': [...], 'labware': [...]}"
|
|
||||||
)
|
|
||||||
|
|
||||||
# 构建工作流图
|
|
||||||
graph = build_protocol_graph(
|
|
||||||
labware_info=labware_info,
|
|
||||||
protocol_steps=protocol_steps,
|
|
||||||
workstation_name=workstation_name,
|
|
||||||
action_resource_mapping=ACTION_RESOURCE_MAPPING,
|
|
||||||
)
|
|
||||||
|
|
||||||
# 校验句柄配置
|
|
||||||
if validate:
|
|
||||||
is_valid, errors = validate_workflow_handles(graph)
|
|
||||||
if not is_valid:
|
|
||||||
import warnings
|
|
||||||
|
|
||||||
for error in errors:
|
|
||||||
warnings.warn(f"句柄校验警告: {error}")
|
|
||||||
|
|
||||||
return graph
|
|
||||||
|
|
||||||
|
|
||||||
def convert_json_to_node_link(
|
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
|
||||||
workstation_name: str = "PRCXi",
|
|
||||||
) -> Dict[str, Any]:
|
|
||||||
"""
|
|
||||||
将 JSON 数据转换为 node-link 格式的字典
|
|
||||||
|
|
||||||
Args:
|
|
||||||
data: JSON 文件路径、字典数据、或 JSON 字符串
|
|
||||||
workstation_name: 工作站名称,默认 "PRCXi"
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
Dict: node-link 格式的工作流数据
|
|
||||||
"""
|
|
||||||
graph = convert_from_json(data, workstation_name)
|
|
||||||
return graph.to_node_link_dict()
|
|
||||||
|
|
||||||
|
|
||||||
def convert_json_to_workflow_list(
|
|
||||||
data: Union[str, PathLike, Dict[str, Any]],
|
|
||||||
workstation_name: str = "PRCXi",
|
|
||||||
) -> List[Dict[str, Any]]:
|
|
||||||
"""
|
|
||||||
将 JSON 数据转换为工作流列表格式
|
|
||||||
|
|
||||||
Args:
|
|
||||||
data: JSON 文件路径、字典数据、或 JSON 字符串
|
|
||||||
workstation_name: 工作站名称,默认 "PRCXi"
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
List: 工作流节点列表
|
|
||||||
"""
|
|
||||||
graph = convert_from_json(data, workstation_name)
|
|
||||||
return graph.to_dict()
|
|
||||||
|
|
||||||
|
|
||||||
# 为了向后兼容,保留下划线前缀的别名
|
|
||||||
_normalize_steps = normalize_steps
|
|
||||||
_normalize_labware = normalize_labware
|
|
||||||
@@ -2,7 +2,7 @@
|
|||||||
<?xml-model href="http://download.ros.org/schema/package_format3.xsd" schematypens="http://www.w3.org/2001/XMLSchema"?>
|
<?xml-model href="http://download.ros.org/schema/package_format3.xsd" schematypens="http://www.w3.org/2001/XMLSchema"?>
|
||||||
<package format="3">
|
<package format="3">
|
||||||
<name>unilabos_msgs</name>
|
<name>unilabos_msgs</name>
|
||||||
<version>0.10.17</version>
|
<version>0.10.15</version>
|
||||||
<description>ROS2 Messages package for unilabos devices</description>
|
<description>ROS2 Messages package for unilabos devices</description>
|
||||||
<maintainer email="changjh@pku.edu.cn">Junhan Chang</maintainer>
|
<maintainer email="changjh@pku.edu.cn">Junhan Chang</maintainer>
|
||||||
<maintainer email="18435084+Xuwznln@users.noreply.github.com">Xuwznln</maintainer>
|
<maintainer email="18435084+Xuwznln@users.noreply.github.com">Xuwznln</maintainer>
|
||||||
|
|||||||
Reference in New Issue
Block a user