Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
P
planning_llh_bgc
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Larkin Heintzman
planning_llh_bgc
Commits
f8a5256a
Commit
f8a5256a
authored
May 01, 2024
by
Bryson Howell
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
initial commit
parent
f8e4a23d
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
302 additions
and
1 deletion
+302
-1
montecarlo.py
mrmh_model/montecarlo.py
+1
-1
test_cpo.py
test_cpo.py
+223
-0
train_cpo.py
train_cpo.py
+78
-0
No files found.
mrmh_model/montecarlo.py
View file @
f8a5256a
...
@@ -254,7 +254,7 @@ class MonteCarlo(params.Default):
...
@@ -254,7 +254,7 @@ class MonteCarlo(params.Default):
self
.
p
=
np
.
zeros
((
self
.
_x_shape
,
self
.
_y_shape
),
dtype
=
float
)
self
.
p
=
np
.
zeros
((
self
.
_x_shape
,
self
.
_y_shape
),
dtype
=
float
)
#Show LP Heatmap
#Show LP Heatmap
for testing
show
=
False
show
=
False
if
(
show
):
if
(
show
):
print
(
'Environment shape =
%
d x
%
d'
%
(
self
.
_x_shape
,
self
.
_y_shape
))
print
(
'Environment shape =
%
d x
%
d'
%
(
self
.
_x_shape
,
self
.
_y_shape
))
...
...
test_cpo.py
0 → 100644
View file @
f8a5256a
from
larrt
import
planning
import
pickle
as
pkl
import
json
from
mrmh_model
import
terrain
,
human
,
montecarlo
,
searcher
import
plotting_utils
as
plotter
import
waypoint_maker
import
matplotlib.pyplot
as
plt
from
scouter.terrain_viewer
import
plot_all
from
matplotlib.backends.backend_pdf
import
PdfPages
from
mrmh_model
import
montecarlo
as
MC
import
numpy
as
np
from
mrmh_model
import
searcher
import
torch
from
gp
import
beziergp
from
gp
import
robotgp
import
sys
from
mrmh_model.params
import
Default
import
datetime
as
dt
import
time
,
os
import
pdb
def
multipage
(
filename
,
figs
=
None
,
dpi
=
200
):
pp
=
PdfPages
(
filename
)
if
figs
is
None
:
figs
=
[
plt
.
figure
(
n
)
for
n
in
plt
.
get_fignums
()]
for
fig
in
figs
:
try
:
fig
.
savefig
(
pp
,
format
=
'pdf'
)
except
:
print
(
'At least one of the figures could not be saved to PDF'
)
pp
.
close
()
def
objective_printer
(
rgp_object
=
None
,
comp_time
=
0.0
,
iteration
=
0
,
folder
=
'blah'
,
time_to_find
=
0
,
find_percentage
=
0
,
save_data
=
False
):
# prints and/or saves the optimization objective values
rgp
=
rgp_object
min_Xstar
=
rgp
.
min_Xstar
[:,
2
]
.
reshape
(
np
.
sqrt
(
rgp
.
min_Xstar
.
shape
[
0
])
.
astype
(
np
.
int
),
-
1
)
.
cpu
()
# find_x = min_Xstar[rgp.mc_handle.find_pt[0], rgp.mc_handle.find_pt[1]]
ystar_mu_min
,
ystar_cov_min
=
rgp
.
eval_batchless
(
rgp
.
min_X
,
rgp
.
min_Xstar
)
ystar_mu_first
,
ystar_cov_first
=
rgp
.
eval_batchless
(
rgp
.
first_X
,
rgp
.
first_Xstar
)
ystar_mu_prev
,
ystar_cov_prev
=
rgp
.
eval_batchless
(
rgp
.
prev_X
,
rgp
.
prev_Xstar
)
ystar_mu_curr
,
ystar_cov_curr
=
rgp
.
eval_batchless
(
rgp
.
X
,
rgp
.
Xstar
)
# undo log odds into just prb
find_mat_min
=
1
-
(
1
/
(
1
+
np
.
exp
(
ystar_mu_min
)))
find_mat_first
=
1
-
(
1
/
(
1
+
np
.
exp
(
ystar_mu_first
)))
find_mat_prev
=
1
-
(
1
/
(
1
+
np
.
exp
(
ystar_mu_prev
)))
find_mat_curr
=
1
-
(
1
/
(
1
+
np
.
exp
(
ystar_mu_curr
)))
risk_cost_min
=
rgp
.
compute_risk_cost_batchless
(
ystar_mu_min
,
ystar_cov_min
)
.
detach
()
.
cpu
()
.
numpy
()
risk_cost_first
=
rgp
.
compute_risk_cost_batchless
(
ystar_mu_first
,
ystar_cov_first
)
.
detach
()
.
cpu
()
.
numpy
()
risk_cost_prev
=
rgp
.
compute_risk_cost_batchless
(
ystar_mu_prev
,
ystar_cov_prev
)
.
detach
()
.
cpu
()
.
numpy
()
risk_cost_curr
=
rgp
.
compute_risk_cost_batchless
(
ystar_mu_curr
,
ystar_cov_curr
)
.
detach
()
.
cpu
()
.
numpy
()
# this only makes sense if paths were optimized
find_ratio_curr
=
(
find_mat_curr
/
find_mat_prev
)
/
((
1
-
find_mat_curr
)
/
(
1
-
find_mat_prev
))
# ratio of successfully finding lp over false positives
find_ratio_total
=
(
find_mat_min
/
find_mat_first
)
/
((
1
-
find_mat_min
)
/
(
1
-
find_mat_first
))
# same thing but compare first and best
# take sample from real lp map
lp_dist
=
rgp
.
mc_handle
.
comp_map
/
np
.
sum
(
rgp
.
mc_handle
.
comp_map
)
# compare chance person is in location and chance we *think* person is in location
search_prb_mat
=
lp_dist
*
find_mat_min
# rgp.min_risk_cost = rgp.risk_cost_data.detach().cpu.numpy().item()
# rgp.min_risk_paths = rgp.robot_points_data
# rgp._min_scaled_risk_cost = rgp.scaled_risk_cost_data
# rgp._min_scaled_path_length_cost = rgp.scaled_path_length_cost_data
# wherever find_ratio is > 1 we have a higher chance of actually finding the person vs getting a false positive (or something like that)
save_dict
=
{
'min_cost_bcls'
:
risk_cost_min
.
item
(),
'curr_cost_bcls'
:
risk_cost_curr
.
item
(),
'curr_delta_bcls'
:
np
.
abs
(
risk_cost_prev
-
risk_cost_curr
),
'init_cost_bcls'
:
risk_cost_first
.
item
(),
'delta_cost_bcls'
:
np
.
abs
(
risk_cost_first
-
risk_cost_min
),
'min_cost_old'
:
rgp
.
min_risk_cost
.
item
(),
'min_risk_old'
:
rgp
.
_min_scaled_risk_cost
.
item
(),
'min_len_old'
:
rgp
.
_min_scaled_path_length_cost
.
item
(),
'find_percentage_curr_old'
:
np
.
count_nonzero
(
find_ratio_curr
>=
1
)
/
np
.
size
(
find_ratio_curr
),
'mean_find_ratio_curr_old'
:
np
.
mean
(
find_ratio_curr
),
'find_percentage_total_old'
:
np
.
count_nonzero
(
find_ratio_total
>=
1
)
/
np
.
size
(
find_ratio_total
),
'mean_find_ratio_total_old'
:
np
.
mean
(
find_ratio_total
),
'search_advantage'
:
np
.
sum
(
search_prb_mat
),
'time_to_find'
:
time_to_find
,
'find_percentage'
:
find_percentage
,
'comp_time'
:
comp_time
}
if
save_data
:
filename
=
dt
.
datetime
.
now
()
.
strftime
(
"
%
d_
%
m_
%
Y-
%
I_
%
M_
%
S_
%
p_"
)
filename
+=
"d-"
+
rgp
.
params
[
'path_style'
]
+
\
"_l-"
+
rgp
.
params
[
'stats_name'
]
+
\
"_i-"
+
str
(
iteration
)
+
".json"
# force create directory
os
.
makedirs
(
'results/'
+
folder
,
exist_ok
=
True
)
with
open
(
'results/'
+
folder
+
'/'
+
filename
,
'w'
)
as
f
:
json
.
dump
(
save_dict
,
f
)
return
save_dict
def
main
(
iteration
=
0
,
parameters
=
-
1
):
# iteration = 1
stime
=
time
.
time
()
params
=
parameters
#MonteCarlo loads LPM, creates searchers (humans on ground)
mc
=
MC
.
MonteCarlo
(
params
=
params
)
# calls terrain builder
mc
.
run_experiment
()
#Run RRT search for paths, and set up optimizer
planner
=
planning
.
Planning
(
params
,
on_terrain
=
mc
.
terrain
,
mode
=
'TOTALDIST'
)
# also calls terrain builder...
rgp
=
robotgp
.
RobotGP
(
mc
,
planner
,
_stime
=
stime
,
parameters
=
params
)
rgp
.
collect_trainData
()
# sets out paths for each robot
#Run ADAM optimizer to curve paths for minimal search risk
rgp
.
optimize_risk_cost_grad
(
_show_detail
=
True
)
robot_paths_local
=
waypoint_maker
.
write_file
(
rgp_object
=
rgp
,
terrain_class
=
mc
.
terrain
,
filename
=
'waypoints.json'
)
# write waypoints to file
etime
=
time
.
time
()
comp_time
=
etime
-
stime
ttf
,
srpc
=
rgp
.
time_to_find
(
robot_paths
=
robot_paths_local
,
searcher_paths
=
mc
.
searcher_class
.
searchers_list
)
print
(
"Total time required: {}"
.
format
(
comp_time
))
objective_printer
(
rgp_object
=
rgp
,
comp_time
=
comp_time
,
iteration
=
iteration
,
folder
=
rgp
.
params
[
'save_folder'
],
save_data
=
rgp
.
params
[
'save_data'
],
time_to_find
=
ttf
,
find_percentage
=
srpc
)
if
rgp
.
params
[
'plot_data'
]:
plot_all
(
parameters
=
params
,
mc_object
=
mc
,
robot_paths
=
robot_paths_local
,
searcher_paths
=
mc
.
searcher_class
.
searchers_list
,
smooth_paths
=
True
,
show_heatmap
=
True
,
show_contours
=
True
,
cs_name
=
'thermal'
)
rgp
.
garbage_cleanup
()
del
rgp
,
planner
,
mc
if
__name__
==
"__main__"
:
#Relative filepaths (Assumes this and ags_grabber projects are in same parent directory)
kentland_heatmap
=
'./LP model/analysis/outputs/ic_1_con_hiker_t12_kentland.csv'
hmpark_heatmap
=
'./LP model/analysis/outputs/ic_1_con_hiker_t12_hmpark.csv'
kentland_linfeat
=
'../ags_grabber/matlab_data/BW_LFandInac_Zelev_kentland.mat'
hmpark_linfeat
=
'../ags_grabber/matlab_data/BW_LFandInac_Zelev_hmpark.mat'
# KENTLAND case
if
True
:
n_max
=
3
#Number of robots
s_max
=
0
#Number of searchers (humans)
h_max
=
0
#Number of humans
#global_fail_max = 1000
global_fail_max
=
5
global_fails
=
0
avg_runs
=
1
start_time
=
time
.
time
()
params
=
({
'lp_model'
:
'stripes'
,
#custom, ring, trust, stripes
'opt_iterations'
:
1000
,
'path_style'
:
'basic'
,
'stats_name'
:
'kentland'
,
'anchor_point'
:
[
37.197730
,
-
80.585233
],
# kentland
'num_searchers'
:
s_max
,
'num_robots'
:
n_max
,
'num_humans'
:
h_max
,
'lp_filename'
:
kentland_heatmap
,
'lin_feat_filename'
:
kentland_linfeat
,
})
params
[
'save_folder'
]
=
'deeprl_{}_n{}_s{}_{}'
.
format
(
params
[
'stats_name'
],
n_max
,
s_max
,
params
[
'path_style'
])
params
=
Default
(
params
)
.
params
counter
=
0
while
counter
<
avg_runs
and
global_fails
<=
global_fail_max
:
# number of averaging runs
print
(
'Running on '
+
torch
.
cuda
.
get_device_name
())
torch
.
cuda
.
empty_cache
()
torch
.
cuda
.
ipc_collect
()
try
:
#try:
# main(iteration = counter, parameters=params)
# print("done run")
#except RuntimeError as e:
# print(e)
# print("\n\n ------ bad memory, re trying ------\n")
# global_fails += 1
# continue
main_cpo
(
iteration
=
counter
,
parameters
=
params
)
#Test DeepRL agent
main_opt
(
iteration
=
counter
,
parameters
=
params
)
#Run path optimizer
print
(
"done run"
)
counter
+=
1
except
AttributeError
as
e
:
print
(
e
)
print
(
"
\n\n
------- bad optimization, re trying ----------
\n
"
)
global_fails
+=
1
#-----------------------------------------------------------------------------------------------
train_cpo.py
0 → 100644
View file @
f8a5256a
from
larrt
import
planning
import
pickle
as
pkl
import
json
from
mrmh_model
import
terrain
,
human
,
montecarlo
,
searcher
import
plotting_utils
as
plotter
import
waypoint_maker
import
matplotlib.pyplot
as
plt
from
scouter.terrain_viewer
import
plot_all
from
matplotlib.backends.backend_pdf
import
PdfPages
from
mrmh_model
import
montecarlo
as
MC
import
numpy
as
np
from
mrmh_model
import
searcher
import
torch
from
gp
import
beziergp
from
gp
import
robotgp
import
sys
from
mrmh_model.params
import
Default
import
datetime
as
dt
import
time
,
os
import
pdb
#Relative filepaths (Assumes this and ags_grabber projects are in same parent directory)
kentland_heatmap
=
'./LP model/analysis/outputs/ic_1_con_hiker_t12_kentland.csv'
hmpark_heatmap
=
'./LP model/analysis/outputs/ic_1_con_hiker_t12_hmpark.csv'
kentland_linfeat
=
'../ags_grabber/matlab_data/BW_LFandInac_Zelev_kentland.mat'
hmpark_linfeat
=
'../ags_grabber/matlab_data/BW_LFandInac_Zelev_hmpark.mat'
def
main
():
#Set up environment parameters
n_max
=
3
#Number of robots
s_max
=
0
#Number of searchers (humans)
h_max
=
0
#Number of humans?
start_time
=
time
.
time
()
#Dictionary of parameters
params
=
({
'lp_model'
:
'stripes'
,
#custom, ring, trust, stripes
'opt_iterations'
:
1000
,
'path_style'
:
'basic'
,
'stats_name'
:
'kentland'
,
'anchor_point'
:
[
37.197730
,
-
80.585233
],
# kentland
'num_searchers'
:
s_max
,
'num_robots'
:
n_max
,
'num_humans'
:
h_max
,
'lp_filename'
:
kentland_heatmap
,
'lin_feat_filename'
:
kentland_linfeat
,
})
params
[
'save_folder'
]
=
'deeprl_{}_n{}_s{}_{}'
.
format
(
params
[
'stats_name'
],
n_max
,
s_max
,
params
[
'path_style'
])
params
=
Default
(
params
)
.
params
#MonteCarlo loads LPM, creates searchers (humans on ground)
mc
=
MC
.
MonteCarlo
(
params
=
params
)
# calls terrain builder
mc
.
run_experiment
()
#Make Gym Environment
if
__name__
==
"__main__"
:
main
()
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment