Go to Post Facebook for FIRST already exists. It's called Facebook :) - dcarr [more]
Home
Go Back   Chief Delphi > Technical > Programming
CD-Media   CD-Spy  
portal register members calendar search Today's Posts Mark Forums Read FAQ rules

 
View Poll Results: What did you use for Vision Processing
Driver Station 28 60.87%
Laptop/Netbook on Robot 3 6.52%
Other 2 4.35%
Did not use/cRIO 13 28.26%
Voters: 46. You may not vote on this poll

Closed Thread
Thread Tools Rate Thread Display Modes
  #1   Spotlight this post!  
Unread 23-08-2012, 10:20
Caboose's Avatar
Caboose Caboose is offline
Programmer - LabVIEW, C++, C#, Java
AKA: James Parks
FRC #0900 (Zebracorns)
Team Role: Alumni
 
Join Date: Jan 2009
Rookie Year: 2008
Location: Seattle, WA
Posts: 72
Caboose has a spectacular aura aboutCaboose has a spectacular aura about
Talking Who used Driver Station for Vision?

I am just wondering who was using their Driver Station or an non cRIO computer for Vision Processing in Rebound Rumble, and what are you going to do next year with the new field data limits.

On a side note, because of the $400 limit of parts on the robot and the new allowance of laptops allowed on robots, would people like a price raise limited for laptops/alt. computing devices that could go on the robot?
__________________
navX Labview Library

"Robots are aluminum shavings, held together by zip-ties."

myManga

Last edited by Caboose : 23-08-2012 at 18:07. Reason: 42
  #2   Spotlight this post!  
Unread 23-08-2012, 11:27
AGPapa's Avatar
AGPapa AGPapa is offline
Registered User
AKA: Antonio Papa
FRC #5895
Team Role: Mentor
 
Join Date: Mar 2012
Rookie Year: 2011
Location: Robbinsville, NJ
Posts: 323
AGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond reputeAGPapa has a reputation beyond repute
Re: Who used Driver Station for Vision?

Where did you hear about new field data limits?
  #3   Spotlight this post!  
Unread 23-08-2012, 11:38
Caboose's Avatar
Caboose Caboose is offline
Programmer - LabVIEW, C++, C#, Java
AKA: James Parks
FRC #0900 (Zebracorns)
Team Role: Alumni
 
Join Date: Jan 2009
Rookie Year: 2008
Location: Seattle, WA
Posts: 72
Caboose has a spectacular aura aboutCaboose has a spectacular aura about
Exclamation Re: Who used Driver Station for Vision?

Quote:
Originally Posted by AGPapa View Post
Where did you hear about new field data limits?
Page 23 of the Einstein Investigation Report: <http://www3.usfirst.org/node/2426>
Quote:
Planned changes to the wireless system to increase robustness were confirmed by the feedback from the wireless
experts consulted as part of the investigation. These items do not directly address failures that occurred on
Einstein but aim to make the wireless network configuration more robust:
...
Quality of Service (QoS) – With a fixed bandwidth cap in place, it becomes critical to prioritize robot
control packets over other types of data such as video. QoS can be used to implement this prioritization
so that robot control packets will continue to flow even if a team exceeds the bandwidth cap with video
or other data.
Ideally each of the 6 teams on the field will get ~50Mbps, most of which will go to FMS/Robot Control data.
__________________
navX Labview Library

"Robots are aluminum shavings, held together by zip-ties."

myManga

Last edited by Caboose : 23-08-2012 at 11:41. Reason: 42
  #4   Spotlight this post!  
Unread 23-08-2012, 13:56
RyanCahoon's Avatar
RyanCahoon RyanCahoon is offline
Disassembling my prior presumptions
FRC #0766 (M-A Bears)
Team Role: Engineer
 
Join Date: Dec 2007
Rookie Year: 2007
Location: Mountain View
Posts: 689
RyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond repute
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by Caboose View Post
Ideally each of the 6 teams on the field will get ~50Mbps, most of which will go to FMS/Robot Control data.
I'm pretty sure the robot control data doesn't take more than about 1KB per update, at around 50 Hz. This gives you 400 Kbps for robot control data.

To transmit a single frame of video with 320x240 resolution and 24bit color is 320*240*24 = 1.8 Mb. Note that the Axis cameras use MJPEG compression as well, so this is a gross overestimate. For targeting purposes, given the network lag that's going to be inherit in the system, you shouldn't need more than 10 fps. Maybe 15 if you want smoother-looking video for display to your drivers.

That's still only 30 Mbps (even with uncompressed video).
__________________
FRC 2046, 2007-2008, Student member
FRC 1708, 2009-2012, College mentor; 2013-2014, Mentor
FRC 766, 2015-, Mentor
  #5   Spotlight this post!  
Unread 23-08-2012, 14:26
Caboose's Avatar
Caboose Caboose is offline
Programmer - LabVIEW, C++, C#, Java
AKA: James Parks
FRC #0900 (Zebracorns)
Team Role: Alumni
 
Join Date: Jan 2009
Rookie Year: 2008
Location: Seattle, WA
Posts: 72
Caboose has a spectacular aura aboutCaboose has a spectacular aura about
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by RyanCahoon View Post
I'm pretty sure the robot control data doesn't take more than about 1KB per update, at around 50 Hz. This gives you 400 Kbps for robot control data.

To transmit a single frame of video with 320x240 resolution and 24bit color is 320*240*24 = 1.8 Mb. Note that the Axis cameras use MJPEG compression as well, so this is a gross overestimate. For targeting purposes, given the network lag that's going to be inherit in the system, you shouldn't need more than 10 fps. Maybe 15 if you want smoother-looking video for display to your drivers.

That's still only 30 Mbps (even with uncompressed video).
Note the word Ideally, 300 Mbps it the theoretical bandwidth for 802.11n with the FMS router and robot in channel bonding mode (this doesn't happen as far as I know), but can drop to under 130 Mbps which would then give each team ~21Mbps or under. Also to do nice detailed image processing my team used 640x480 resolution on two cameras in stereo, pushing us to ~14 Megabits(7Mb*2) for both images at ~15FPS, excluding network overhead, this according to the FMS people in Raleigh was getting close to taxing the FMS when we were the only team using vision on the field during a match.
__________________
navX Labview Library

"Robots are aluminum shavings, held together by zip-ties."

myManga
  #6   Spotlight this post!  
Unread 23-08-2012, 14:29
Joe Ross's Avatar Unsung FIRST Hero
Joe Ross Joe Ross is offline
Registered User
FRC #0330 (Beachbots)
Team Role: Engineer
 
Join Date: Jun 2001
Rookie Year: 1997
Location: Los Angeles, CA
Posts: 8,600
Joe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond reputeJoe Ross has a reputation beyond repute
Re: Who used Driver Station for Vision?

Here's is what the Q/A said for last year:

Quote:
There are currently no bandwidth limits in place in the field network. In theory, each team has 50Mbits/second (300Mbits/6) available, but that’s not actually realistic. In reality, each team is likely to have ~10-12Mbits/s available. This rate will vary depending on the location of the radio on the Robot and the amount of wireless traffic present in the venue at 5GHz. While this information may help give teams an idea of what to expect, note that there is no guaranteed level of bandwidth on the playing field.
  #7   Spotlight this post!  
Unread 23-08-2012, 15:54
RyanCahoon's Avatar
RyanCahoon RyanCahoon is offline
Disassembling my prior presumptions
FRC #0766 (M-A Bears)
Team Role: Engineer
 
Join Date: Dec 2007
Rookie Year: 2007
Location: Mountain View
Posts: 689
RyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond repute
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by Caboose View Post
Note the word Ideally, 300 Mbps it the theoretical bandwidth for 802.11n with the FMS router and robot in channel bonding mode (this doesn't happen as far as I know), but can drop to under 130 Mbps which would then give each team ~21Mbps or under.
Acknowledged - just pointing out that the needed bandwidth is fairly close to the actually available bandwidth, not many orders of magnitude difference. I'm actually hoping they can implement a robust QOS system; it will make our jobs much easier instead of having to implement throttling ourselves.

Quote:
Originally Posted by Caboose View Post
Also to do nice detailed image processing my team used 640x480 resolution on two cameras in stereo, pushing us to ~14 Megabits(7Mb*2) for both images at ~15FPS
Did you do any testing to compare your accuracy against a lower resolution?
__________________
FRC 2046, 2007-2008, Student member
FRC 1708, 2009-2012, College mentor; 2013-2014, Mentor
FRC 766, 2015-, Mentor
  #8   Spotlight this post!  
Unread 23-08-2012, 16:06
R.C.'s Avatar
R.C. R.C. is offline
2017... Oooh Kill em, Swerve!
AKA: Owner, WestCoast Products
FRC #1323 (MadTown Robotics)
Team Role: Engineer
 
Join Date: Feb 2008
Rookie Year: 2006
Location: Madera, CA
Posts: 2,188
R.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond reputeR.C. has a reputation beyond repute
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by Caboose View Post
I am just wondering who was using their Driver Station or an non cRIO computer for Vision Processing in Rebound Rumble, and what are you going to do next year with the new field data limits.
341 used off board vision using Java/SmartDashboard. It worked really well for them. One of the top 5 teams in FRC this year. We used their code as well, it made a huge difference for us.

They were kind enough to put their code up on Delphi, so take a look if your interested.

-RC
__________________
R.C.
Owner, WestCoast Products || Twitter
MadTown Robotics Team 1323
  #9   Spotlight this post!  
Unread 23-08-2012, 18:02
Caboose's Avatar
Caboose Caboose is offline
Programmer - LabVIEW, C++, C#, Java
AKA: James Parks
FRC #0900 (Zebracorns)
Team Role: Alumni
 
Join Date: Jan 2009
Rookie Year: 2008
Location: Seattle, WA
Posts: 72
Caboose has a spectacular aura aboutCaboose has a spectacular aura about
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by RyanCahoon View Post
Did you do any testing to compare your accuracy against a lower resolution?
Yes, we tested on lower resolution. Refer to my teams white paper for examples of a 640x480 image and how far apart our cameras were. The reason for the large images is for calculations, low resolution images give larger errors in our distance and angle calculations due to the pixels.


*Edited top Question*
__________________
navX Labview Library

"Robots are aluminum shavings, held together by zip-ties."

myManga

Last edited by Caboose : 23-08-2012 at 18:07. Reason: 42
  #10   Spotlight this post!  
Unread 23-08-2012, 18:17
RyanCahoon's Avatar
RyanCahoon RyanCahoon is offline
Disassembling my prior presumptions
FRC #0766 (M-A Bears)
Team Role: Engineer
 
Join Date: Dec 2007
Rookie Year: 2007
Location: Mountain View
Posts: 689
RyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond repute
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by Caboose View Post
Yes, we tested on lower resolution. [...] The reason for the large images is for calculations, low resolution images give larger errors in our distance and angle calculations due to the pixels.
That's what I had assumed, but I'm curious if you noted how much larger the errors were. I'd also be interested in the errors of the stereo distance measurement vs the perspective distance measurement as a function of distance, if you have them.
__________________
FRC 2046, 2007-2008, Student member
FRC 1708, 2009-2012, College mentor; 2013-2014, Mentor
FRC 766, 2015-, Mentor
  #11   Spotlight this post!  
Unread 23-08-2012, 19:29
F22Rapture's Avatar
F22Rapture F22Rapture is offline
College Student, Mentor
AKA: Daniel A
FRC #3737 (4H Rotoraptors)
Team Role: Mentor
 
Join Date: Jan 2012
Rookie Year: 2012
Location: Goldsboro, NC
Posts: 476
F22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant futureF22Rapture has a brilliant future
Re: Who used Driver Station for Vision?

Quote:
Originally Posted by Joe Ross View Post
Here's is what the Q/A said for last year:
Due to the events on Einstein some of the rules regarding networks have changed. One of those is that there's going to be a bandwidth limit for each robot and QoS to ensure the driver commands can make it through with as little latency as possible.
  #12   Spotlight this post!  
Unread 23-08-2012, 20:43
apalrd's Avatar
apalrd apalrd is offline
More Torque!
AKA: Andrew Palardy (Most people call me Palardy)
VRC #3333
Team Role: College Student
 
Join Date: Mar 2009
Rookie Year: 2009
Location: Auburn Hills, MI
Posts: 1,347
apalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond reputeapalrd has a reputation beyond repute
Re: Who used Driver Station for Vision?

We ran our vision on Dashboard.

I set the fps limit to 20fps. Due to the design of the control algorithm we used, more frames is better, so we try for 20. We get images at 320x240 (we determined this was the smallest size that achieved adequate resolution with the fish-eye lens we used).

If we encounter any bandwidth limit in the future, we'll either run smaller or slower. Running at 320x240 at 10fps is still far better than what the cRio on its own is capable of (I don't think we could get more than 6fps at any resolution, given the amount of processor utilized by non-vision code), and any solution that is weight neutral on the robot the solution we will almost always choose.
__________________
Kettering University - Computer Engineering
Kettering Motorsports
Williams International - Commercial Engines - Controls and Accessories
FRC 33 - The Killer Bees - 2009-2012 Student, 2013-2014 Advisor
VEX IQ 3333 - The Bumble Bees - 2014+ Mentor

"Sometimes, the elegant implementation is a function. Not a method. Not a class. Not a framework. Just a function." ~ John Carmack
  #13   Spotlight this post!  
Unread 24-08-2012, 17:12
Caboose's Avatar
Caboose Caboose is offline
Programmer - LabVIEW, C++, C#, Java
AKA: James Parks
FRC #0900 (Zebracorns)
Team Role: Alumni
 
Join Date: Jan 2009
Rookie Year: 2008
Location: Seattle, WA
Posts: 72
Caboose has a spectacular aura aboutCaboose has a spectacular aura about
Lightbulb Re: Who used Driver Station for Vision?

Quote:
Originally Posted by apalrd View Post
If we encounter any bandwidth limit in the future, we'll either run smaller or slower.
I would think a better plan would be to have the vision processing calculated on the robot, use of cheaper and better USB cameras with low image capture lag, with a light and powerful laptop and send relevant data to the Driver Station/Robot. But alas there is a dreaded $400 limit to ALL parts, if only laptops that would go on the robot could be a little more expensive the FMS would not have a need to worry about large images clogging the field network...
__________________
navX Labview Library

"Robots are aluminum shavings, held together by zip-ties."

myManga
  #14   Spotlight this post!  
Unread 24-08-2012, 17:27
RyanCahoon's Avatar
RyanCahoon RyanCahoon is offline
Disassembling my prior presumptions
FRC #0766 (M-A Bears)
Team Role: Engineer
 
Join Date: Dec 2007
Rookie Year: 2007
Location: Mountain View
Posts: 689
RyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond reputeRyanCahoon has a reputation beyond repute
Re: Who used Driver Station for Vision?

If you're worried about the price limit and weight, you shouldn't even be looking at laptops. Build your own Mini-ITX board. You can buy/bill the components separately and to your needed specifications, and that way you don't have to carry around the screen. There are power supplies on that site designed for car computers that are tolerant to voltage drop-out as well, so you can even power the computer off of the robot's battery, saving you more weight.
__________________
FRC 2046, 2007-2008, Student member
FRC 1708, 2009-2012, College mentor; 2013-2014, Mentor
FRC 766, 2015-, Mentor
  #15   Spotlight this post!  
Unread 24-08-2012, 18:45
cbale2000's Avatar
cbale2000 cbale2000 is offline
Registered User
AKA: Chris Bale
FRC #5712 (Gray Matter)
Team Role: Mentor
 
Join Date: Apr 2006
Rookie Year: 2004
Location: Saginaw, MI
Posts: 970
cbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond reputecbale2000 has a reputation beyond repute
Re: Who used Driver Station for Vision?

For me the bigger question is, why doesn't FIRST just ditch the cRio and go with laptop controlled robots?

All it would take is a USB motor controller + I/O board (and since we already have FIRST-specific electronics that ship with the cRio, I don't see this as being a huge issue) and almost any laptop and you have a control system that can be easily upgraded/updated, has an independent backup battery (no worries about power loss on your controller), can be programmed in virtually any language, can run vision processing without eating network bandwidth, and, depending on what you get, can be cheaper and have more features than the cRio.
__________________
2005-2008: FRC 703 Phoenix - Driver
2009-2016: FRC 703 Phoenix - Mentor
2017-????: FRC 5712 Gray Matter - Mentor
Closed Thread


Thread Tools
Display Modes Rate This Thread
Rate This Thread:

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

vB code is On
Smilies are On
[IMG] code is On
HTML code is Off
Forum Jump


All times are GMT -5. The time now is 01:36.

The Chief Delphi Forums are sponsored by Innovation First International, Inc.


Powered by vBulletin® Version 3.6.4
Copyright ©2000 - 2017, Jelsoft Enterprises Ltd.
Copyright © Chief Delphi