#!/usr/bin/env python3 ''' NAME: sta_connect2.py Purpose: Test will create a station, create TCP and UDP traffic, run it a short amount of time, and verify whether traffic was sent and received. It also verifies the station connected to the requested BSSID if bssid is specified as an argument. The script will clean up the station and connections at the end of the test. EXAMPLE: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth1 --radio wiphy1 CLI Example for kpi.csv report output: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth2 --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234 CLI Example for kpi.csv, variable tx/rx rates, and pdu size: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth2 --download_bps 768000 --upload_bps 256000 --side_a_pdu 300 --side_b_pdu 750 --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234 COPYRIGHT: Copyright 2021 Candela Technologies Inc License: Free to distribute and modify. LANforge systems must be licensed. ''' import sys import os import csv import importlib import argparse from pprint import pformat import time import logging import datetime import platform if sys.version_info[0] != 3: print("This script requires Python 3") exit(1) sys.path.append(os.path.join(os.path.abspath(__file__ + "../../../"))) LFUtils = importlib.import_module("py-json.LANforge.LFUtils") removeCX = LFUtils.removeCX removeEndps = LFUtils.removeEndps realm = importlib.import_module("py-json.realm") Realm = realm.Realm influx = importlib.import_module("py-scripts.influx_utils") RecordInflux = influx.RecordInflux lf_report = importlib.import_module("py-scripts.lf_report") lf_graph = importlib.import_module("py-scripts.lf_graph") lf_kpi_csv = importlib.import_module("py-scripts.lf_kpi_csv") logger = logging.getLogger(__name__) lf_logger_config = importlib.import_module("py-scripts.lf_logger_config") OPEN = "open" WEP = "wep" WPA = "wpa" WPA2 = "wpa2" WPA3 = "wpa3" OWE = "owe" MODE_AUTO = 0 class StaConnect2(Realm): def __init__(self, host, port, _dut_ssid="jedway-open-1", _dut_passwd="NA", _dut_bssid="", _user="", _passwd="", _radio="wiphy0", _sta_mode=0, _influx_host=None, _influx_db=None, _influx_user=None, _influx_passwd=None, _resource=1, _upstream_resource=1, _upstream_port="eth1", _sta_name=None, _sta_prefix=None, _bringup_time_sec=300, debug_=False, _dut_security=OPEN, _exit_on_error=False, _cleanup_on_exit=True, _clean_all_sta=False, _runtime_sec=None, kpi_csv=None, outfile=None, download_bps=None, upload_bps=None, side_a_pdu=None, side_b_pdu=None, _exit_on_fail=False): # do not use `super(LFCLiBase,self).__init__(self, host, port, _debugOn) # that is py2 era syntax and will force self into the host variable, making you # very confused. super().__init__(host, port, debug_=debug_, _exit_on_fail=_exit_on_fail) self.host = host self.port = port self.debug = debug_ self.dut_security = _dut_security self.dut_ssid = _dut_ssid self.dut_passwd = _dut_passwd self.dut_bssid = _dut_bssid self.user = _user self.passwd = _passwd self.sta_mode = _sta_mode # See add_sta LANforge CLI users guide entry self.radio = _radio self.resource = _resource self.outfile = outfile self.kpi_csv = kpi_csv self.download_bps = download_bps self.upload_bps = upload_bps self.side_a_pdu = side_a_pdu self.side_b_pdu = side_b_pdu self.epoch_time = int(time.time()) self.upstream_resource = _upstream_resource self.upstream_port = _upstream_port self.runtime_secs = _runtime_sec self.cleanup_on_exit = _cleanup_on_exit self.clean_all_sta = _clean_all_sta self.sta_url_map = None # defer construction self.upstream_url = None # defer construction self.station_names = [] if _sta_name: self.station_names = [_sta_name] self.sta_prefix = _sta_prefix self.bringup_time_sec = _bringup_time_sec self.resulting_stations = {} self.resulting_endpoints = {} self.station_profile = None self.l3_udp_profile = None self.l3_tcp_profile = None self.influx_host = _influx_host self.influx_db = _influx_db self.influx_user = _influx_user self.influx_passwd = _influx_passwd self.name_prefix = "tcp" self.use_existing_sta = False self.cx_profile = self.new_l3_cx_profile() self.cx_profile.host = self.host self.cx_profile.port = self.port self.cx_profile.name_prefix = self.name_prefix if self.outfile is not None: results = self.outfile[:-4] results = results + "-results.csv" self.csv_results_file = open(results, "w") self.csv_results_writer = csv.writer(self.csv_results_file, delimiter=",") if self.sta_prefix is None: self.sta_prefix = "1.%s.sta" % (self.resource) def get_kpi_results(self): # make json call to get kpi results endp_list = self.json_get( "endp?fields=name,eid,delay,jitter,rx+rate,rx+rate+ll,rx+bytes,rx+drop+%25,rx+pkts+ll", debug_=False) logger.info("endp_list: {endp_list}".format(endp_list=endp_list)) def get_csv_name(self): logger.info("self.csv_results_file {}".format(self.csv_results_file.name)) return self.csv_results_file.name # Query all endpoints to generate rx and other stats, returned # as an array of objects. def get_rx_values(self): endp_list = self.json_get( "endp?fields=name,eid,delay,jitter,rx+rate,rx+rate+ll,rx+bytes,rx+drop+%25,rx+pkts+ll", debug_=True) # logger.info("endp_list: {endp_list}".format(endp_list=endp_list)) endp_rx_drop_map = {} endp_rx_map = {} our_endps = {} endps = [] total_ul = 0 total_ul_ll = 0 total_dl = 0 total_dl_ll = 0 udp_dl = 0 tcp_dl = 0 udp_ul = 0 tcp_ul = 0 ''' for e in self.cx_profile.created_endp.keys(): our_endps[e] = e print("our_endps {our_endps}".format(our_endps=our_endps)) ''' for endp_name in endp_list['endpoint']: if endp_name != 'uri' and endp_name != 'handler': for item, endp_value in endp_name.items(): # if item in our_endps: if True: endps.append(endp_value) logger.debug("endpoint: {item} value:\n".format(item=item)) logger.debug(endp_value) # print("item {item}".format(item=item)) for value_name, value in endp_value.items(): if value_name == 'rx bytes': endp_rx_map[item] = value if value_name == 'rx rate': endp_rx_map[item] = value # print("value {value}".format(value=value)) if value_name == 'rx rate ll': endp_rx_map[item] = value # print("value {value}".format(value=value)) if value_name == 'rx pkts ll': endp_rx_map[item] = value if value_name == 'rx drop %': endp_rx_drop_map[item] = value if value_name == 'rx rate': # This hack breaks for mcast or if someone names endpoints weirdly. # logger.info("item: ", item, " rx-bps: ", value_rx_bps) # info for download test data if item.startswith("udp") and item.endswith("-A"): udp_ul += int(value) # logger.info(udp_udl) elif item.startswith("tcp") and item.endswith("-A"): tcp_ul += int(value) # logger.info(tcp_dl) # info for upload test data if item.startswith("udp") and item.endswith("-B"): udp_dl += int(value) # logger.info(udp_ul) elif item.startswith("tcp") and item.endswith("-B"): tcp_dl += int(value) # logger.info(tcp_ul) # combine total download (tcp&udp) and upload (tcp&udp) test data if item.endswith("-A"): total_dl += int(value) # logger.info(total_dl) else: total_ul += int(value) if value_name == 'rx rate ll': # This hack breaks for mcast or if someone # names endpoints weirdly. if item.endswith("-A"): total_dl_ll += int(value) else: total_ul_ll += int(value) # logger.debug("total-dl: ", total_dl, " total-ul: ", total_ul, "\n") return endp_rx_map, endp_rx_drop_map, endps, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul, total_dl_ll, total_ul_ll # Common code to generate timestamp for CSV files. def time_stamp(self): return time.strftime('%m_%d_%Y_%H_%M_%S', time.localtime(self.epoch_time)) def get_station_url(self, sta_name_=None): if sta_name_ is None: raise ValueError("get_station_url wants a station name") if self.sta_url_map is None: self.sta_url_map = "port/1/%s/%s" % (self.resource, sta_name_) logger.info("self.sta_url: %s", self.sta_url_map) return self.sta_url_map ''' if self.sta_url_map is None: self.sta_url_map = {} for sta_name in self.station_names: split_sta_name = sta_name.split('.') only_sta_name = split_sta_name[2] # sta_url = self.get_station_url(only_sta_name) self.sta_url_map[only_sta_name] = "port/1/%s/%s" % (self.resource, only_sta_name) # return self.sta_url_map[sta_name_] return self.sta_url_map ''' def get_upstream_url(self): if self.upstream_url is None: self.upstream_url = "port/1/%s/%s" % (self.upstream_resource, self.upstream_port) return self.upstream_url # Compare pre-test values to post-test values def compare_vals(self, name, post_val, print_pass=False, print_fail=True): # print(f"Comparing {name}") # logger.info(name) # logger.info(postVal) if post_val > 0: pass_msg = str(post_val) + " bps" self._pass("%s %s" % (name, pass_msg), print_pass) else: fail_msg = str(post_val) + " bps" self._fail("%s did not report traffic: %s" % (name, fail_msg), print_fail) def remove_stations(self): for name in self.station_names: LFUtils.removePort(self.resource, name, self.lfclient_url) def num_associated(self, bssid): counter = 0 # print("there are %d results" % len(self.station_results)) fields = "_links,port,alias,ip,ap,port+type" self.station_results = self.find_ports_like("%s*" % self.sta_prefix, fields, debug_=False) logger.info("num_associated() - self.station_results: {}".format(self.station_results)) if (self.station_results is None) or (len(self.station_results) < 1): self.get_failed_result_list() for eid, record in self.station_results.items(): # logger.info("-- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- ") logger.debug(pformat(eid)) # logger.debug(pformat(record)) if record["ap"] == bssid: counter += 1 # logger.info("-- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- -- ") return counter def clear_test_results(self): self.resulting_stations = {} self.resulting_endpoints = {} super().clear_test_results() # super(StaConnect, self).clear_test_results().test_results.clear() def setup(self, extra_securities=None): if extra_securities is None: extra_securities = [] self.clear_test_results() self.check_connect() upstream_json = self.json_get("%s?fields=alias,phantom,down,port,ip" % self.get_upstream_url(), debug_=False) if upstream_json is None: self._fail(message="Unable to query %s, bye" % self.upstream_port, print_=True) return False if upstream_json['interface']['ip'] == "0.0.0.0": if self.debug: logger.debug(pformat(upstream_json)) self._fail("Warning: %s lacks ip address" % self.get_upstream_url(), print_=True) return False self.pre_cleanup() # Create stations and turn dhcp on self.station_profile = self.new_station_profile() self.station_profile.sta_prefix = self.sta_prefix self.station_profile.mode = self.sta_mode if self.dut_security == WPA2: self.station_profile.use_security(security_type="wpa2", ssid=self.dut_ssid, passwd=self.dut_passwd) elif self.dut_security == OPEN: self.station_profile.use_security(security_type="open", ssid=self.dut_ssid, passwd="[BLANK]") elif self.dut_security == WPA: self.station_profile.use_security(security_type="wpa", ssid=self.dut_ssid, passwd=self.dut_passwd) elif self.dut_security == WEP: self.station_profile.use_security(security_type="wep", ssid=self.dut_ssid, passwd=self.dut_passwd) elif self.dut_security == WPA3: self.station_profile.use_security(security_type="wpa3", ssid=self.dut_ssid, passwd=self.dut_passwd) elif self.dut_security == OWE: self.station_profile.use_security(security_type="owe", ssid=self.dut_ssid, passwd="[BLANK]") self.station_profile.set_command_flag("add_sta", "use-owe", 1) self.station_profile.set_command_flag("add_sta", "8021x_radius", 1) self.station_profile.set_command_param("add_sta", "ieee80211w", 2) self.station_profile.set_wifi_extra(key_mgmt="OWE") else: raise ValueError(f"unknown setting for station security: {self.dut_security}") self.station_profile.set_command_flag("add_sta", "create_admin_down", 1) for security in extra_securities: self.station_profile.add_security_extra(security=security) # print("Adding new stations ", end="") logger.info("Adding new stations ") self.station_profile.create(radio=self.radio, sta_names_=self.station_names, up_=False, debug=self.debug, suppress_related_commands_=True) LFUtils.wait_until_ports_appear(self.lfclient_url, self.station_names, debug=self.debug) # Create UDP endpoints self.l3_udp_profile = self.new_l3_cx_profile() ''' self.l3_udp_profile.side_a_min_bps = 128000 self.l3_udp_profile.side_b_min_bps = 128000 self.l3_udp_profile.side_a_min_pdu = 1200 self.l3_udp_profile.side_b_min_pdu = 1500 ''' self.l3_udp_profile.side_a_min_bps = self.download_bps self.l3_udp_profile.side_b_min_bps = self.upload_bps self.l3_udp_profile.side_a_min_pdu = self.side_a_pdu self.l3_udp_profile.side_b_min_pdu = self.side_b_pdu self.l3_udp_profile.report_timer = 1000 self.l3_udp_profile.name_prefix = "udp" port_list = list(self.find_ports_like("%s+" % self.sta_prefix)) if (port_list is None) or (len(port_list) < 1): raise ValueError("Unable to find ports named '%s'+" % self.sta_prefix) self.l3_udp_profile.create(endp_type="lf_udp", side_a=port_list, side_b="%s.%s" % (int(self.upstream_resource), self.upstream_port), suppress_related_commands=True) # Create TCP endpoints self.l3_tcp_profile = self.new_l3_cx_profile() ''' self.l3_tcp_profile.side_a_min_bps = 128000 self.l3_tcp_profile.side_b_min_bps = 56000 ''' self.l3_tcp_profile.side_a_min_bps = self.download_bps self.l3_tcp_profile.side_b_min_bps = self.upload_bps self.l3_tcp_profile.side_a_min_pdu = self.side_a_pdu self.l3_tcp_profile.side_b_min_pdu = self.side_b_pdu self.l3_tcp_profile.name_prefix = self.name_prefix self.l3_tcp_profile.report_timer = 1000 self.l3_tcp_profile.create(endp_type="lf_tcp", side_a=list(self.find_ports_like("%s+" % self.sta_prefix)), side_b="%s.%s" % (int(self.upstream_resource), self.upstream_port), suppress_related_commands=True) def start(self): if self.station_profile is None: self._fail("Incorrect setup") logger.debug(pformat(self.station_profile)) if self.station_profile.up is None: self._fail("Incorrect station profile, missing profile.up") if not self.station_profile.up: logger.info("Bringing ports up...") # logger.info("start() - self.lf_client_url: %s", self.lfclient_url) data = {"shelf": 1, "resource": self.resource, "port": "ALL", "probe_flags": 1} # logger.info("start() - data1: %s", data) self.json_post("/cli-json/nc_show_ports", data) # print("start() - ln 252 - self.j") self.station_profile.admin_up() LFUtils.waitUntilPortsAdminUp(self.resource, self.lfclient_url, self.station_names) self.csv_add_column_headers() if self.influx_db: grapher = RecordInflux(_influx_host=self.influx_host, _influx_db=self.influx_db, _influx_user=self.influx_user, _influx_passwd=self.influx_passwd, _longevity=1, _devices=self.station_names, _monitor_interval=1, _target_kpi=['bps rx']) # station_info = self.jsonGet(self.mgr_url, "%s?fields=port,ip,ap" % (self.getStaUrl())) duration = 0 maxTime = self.bringup_time_sec ip = "0.0.0.0" ap = "" # print("Waiting for %s stations to associate to AP: " % len(self.station_names), end="") logger.info("Waiting for %s stations to associate to AP: ", len(self.station_names)) connected_stations = {} while (len(connected_stations.keys()) < len(self.station_names)) and (duration < maxTime): duration += 3 time.sleep(3) print(".", end="") for sta_name in self.station_names: split_sta_name = sta_name.split('.') # logger.info("sta_name.split: %s", split_sta_name) only_sta_name = split_sta_name[2] sta_url = self.get_station_url(only_sta_name) station_info = self.json_get(sta_url + "?fields=port,ip,ap") # logger.info("start() - sta_url: %s", sta_url) # logger.info("start() - station_info: %s", station_info) # LFUtils.debug_printer.pprint(station_info) if (station_info) and ("interface" in station_info): if "ip" in station_info["interface"]: ip = station_info["interface"]["ip"] if "ap" in station_info["interface"]: ap = station_info["interface"]["ap"] if (ap == "Not-Associated") or (ap == ""): if self.debug: # print(" -%s," % sta_name, end="") logger.info(" -%s,", sta_name) else: if ip == "0.0.0.0": if self.debug: # print(" %s (0.0.0.0)" % sta_name, end="") logger.info(" %s (0.0.0.0)", sta_name) else: connected_stations[sta_name] = sta_url data = { "shelf": 1, "resource": self.resource, "port": "ALL", "probe_flags": 1 } # logger.info("start() - data2: %s", data) self.json_post("/cli-json/nc_show_ports", data) if self.influx_db: grapher.getdata() # LFUtils.wait_until_ports_appear(port_list=self.station_names, debug=self.debug) LFUtils.wait_until_ports_appear(self.lfclient_url, port_list=self.station_names, debug=self.debug) for sta_name in self.station_names: sta_url = self.get_station_url(sta_name) station_info = self.json_get(sta_url) # + "?fields=port,ip,ap") logger.info(f"station_info:{station_info}") if station_info is None: logger.info("unable to query %s", sta_url) self.resulting_stations[sta_url] = station_info try: ap = station_info["interface"]["ap"] except Exception as e: logger.info(f"Exception error:{e}") logger.info(e) ap = "NULL" logger.info(f"ap value:{ap}") ip = station_info["interface"]["ip"] if (ap == "") or (ap == "Not-Associated"): time.sleep(10) station_info = self.json_get(sta_url) logger.info(f"station_info after 10 seconds:{station_info}") ap = station_info["interface"]["ap"] logger.info(f"ap value after 10 seconds:{ap}") if (ap != "") and (ap != "Not-Associated"): print(" %s +AP %s, " % (sta_name, ap), end="") if self.dut_bssid != "": if self.dut_bssid.lower() == ap.lower(): self._pass(sta_name + " connected to BSSID: " + ap) # self.test_results.append("PASSED: ) logger.info("PASSED: Connected to BSSID: %s", ap) else: self._fail( "%s connected to wrong BSSID, requested: %s Actual: %s" % (sta_name, self.dut_bssid, ap)) else: logger.info(f"{sta_name} did not connect to AP") self._fail(sta_name + " did not connect to AP") return False if ip == "0.0.0.0": self._fail("%s did not get an ip. Ending test" % sta_name) else: self._pass("%s connected to AP: %s With IP: %s" % (sta_name, ap, ip)) if not self.passes(): if self.cleanup_on_exit: logger.info("Cleaning up...") self.remove_stations() return False # start cx traffic logger.info("Starting CX Traffic") self.l3_udp_profile.start_cx() self.l3_tcp_profile.start_cx() time.sleep(1) # Refresh stats self.l3_udp_profile.refresh_cx() self.l3_tcp_profile.refresh_cx() def collect_endp_stats(self, endp_map): logger.info("Collecting Data") fields = "/all" for (cx_name, endps) in endp_map.items(): try: endp_url = "/endp/%s%s" % (endps[0], fields) endp_json = self.json_get(endp_url) self.resulting_endpoints[endp_url] = endp_json ptest_a_tx = endp_json['endpoint']['tx bytes'] ptest_a_rx = endp_json['endpoint']['rx bytes'] # ptest = self.json_get("/endp/%s?fields=tx+bytes,rx+bytes" % cx_names[cx_name]["b"]) endp_url = "/endp/%s%s" % (endps[1], fields) endp_json = self.json_get(endp_url) self.resulting_endpoints[endp_url] = endp_json ptest_b_tx = endp_json['endpoint']['tx bytes'] ptest_b_rx = endp_json['endpoint']['rx bytes'] self.compare_vals(endps[0], ptest_a_tx) self.compare_vals(endps[0], ptest_a_rx) self.compare_vals(endps[1], ptest_b_tx) self.compare_vals(endps[1], ptest_b_rx) except Exception as e: self.error(e) def stop(self): # stop cx traffic logger.info("Stopping CX Traffic") self.l3_udp_profile.stop_cx() self.l3_tcp_profile.stop_cx() # Refresh stats logger.info("Refresh CX stats") self.l3_udp_profile.refresh_cx() self.l3_tcp_profile.refresh_cx() logger.info("Sleeping for 5 seconds") time.sleep(5) # get data for endpoints JSON self.collect_endp_stats(self.l3_udp_profile.created_cx) self.collect_endp_stats(self.l3_tcp_profile.created_cx) # print("self.l3_udp_profile.created_cx -ln 412-: ", self.l3_udp_profile.created_cx) # print("self.l3_tcp_profile.created_cx: ", self.l3_tcp_profile.created_cx) def cleanup(self): # remove all endpoints and cxs if self.cleanup_on_exit: for sta_name in self.station_names: split_sta_name = sta_name.split('.') # logger.info("sta_name.split: %s", split_sta_name) only_sta_name = split_sta_name[2] LFUtils.removePort(self.resource, only_sta_name, self.lfclient_url) curr_endp_names = [] removeCX(self.lfclient_url, self.l3_udp_profile.get_cx_names()) removeCX(self.lfclient_url, self.l3_tcp_profile.get_cx_names()) for (cx_name, endp_names) in self.l3_udp_profile.created_cx.items(): curr_endp_names.append(endp_names[0]) curr_endp_names.append(endp_names[1]) for (cx_name, endp_names) in self.l3_tcp_profile.created_cx.items(): curr_endp_names.append(endp_names[0]) curr_endp_names.append(endp_names[1]) removeEndps(self.lfclient_url, curr_endp_names, debug=self.debug) def pre_cleanup(self): self.cx_profile.cleanup_prefix() # do not clean up station if existed prior to test if not self.use_existing_sta: for sta in self.station_names: self.rm_port(sta, check_exists=True, debug_=self.debug) # builds test data into kpi.csv report def record_kpi_csv( self, sta_list, total_test, total_pass, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul): ''' logger.debug( "NOTE: Adding kpi to kpi.csv, sta_count {sta_list} total-download-bps:{total_dl} upload: {total_ul} bi-directional: {total}\n".format( sta_list=sta_list, total_dl=total_dl, total_ul=total_ul, total=(total_ul + total_dl))) logger.debug( "NOTE: Adding kpi to kpi.csv, sta_count {sta_list} total-download-bps:{total_dl_ll_bps} upload: {total_ul_ll_bps} bi-directional: {total_ll}\n".format( sta_list=sta_list, total_dl_ll_bps=total_dl_ll_bps, total_ul_ll_bps=total_ul_ll_bps, total_ll=(total_ul_ll_bps + total_dl_ll_bps))) ''' # the short description will allow for more data to show up in one test-tag graph # logic for Subtest-Pass & Subtest-Fail columns subpass_udp_dl = 0 subpass_udp_ul = 0 subpass_tcp_dl = 0 subpass_tcp_ul = 0 subfail_udp_dl = 1 subfail_udp_ul = 1 subfail_tcp_dl = 1 subfail_tcp_ul = 1 if udp_dl > 0: subpass_udp_dl = 1 subfail_udp_dl = 0 if udp_ul > 0: subpass_udp_ul = 1 subfail_udp_ul = 0 if tcp_dl > 0: subpass_tcp_dl = 1 subfail_tcp_dl = 0 if tcp_ul > 0: subpass_tcp_ul = 1 subfail_tcp_ul = 0 # logic for pass/fail column # total_test & total_pass values from lfcli_base.py if total_test == total_pass: pass_fail = "PASS" else: pass_fail = "FAIL" # kpi data for UDP download traffic results_dict = self.kpi_csv.kpi_csv_get_dict_update_time() results_dict['Graph-Group'] = "UDP Download Rate" results_dict['pass/fail'] = pass_fail results_dict['Subtest-Pass'] = subpass_udp_dl results_dict['Subtest-Fail'] = subfail_udp_dl results_dict['short-description'] = "UDP-DL {download_bps} bps pdu {side_a_pdu} {sta_list} STA".format( download_bps=self.download_bps, side_a_pdu=self.side_a_pdu, sta_list=sta_list) results_dict['numeric-score'] = "{}".format(udp_dl) results_dict['Units'] = "bps" self.kpi_csv.kpi_csv_write_dict(results_dict) # kpi data for UDP upload traffic results_dict['Graph-Group'] = "UDP Upload Rate" results_dict['pass/fail'] = pass_fail results_dict['Subtest-Pass'] = subpass_udp_ul results_dict['Subtest-Fail'] = subfail_udp_ul results_dict['short-description'] = "UDP-UL {upload_bps} bps pdu {side_b_pdu} {sta_list} STA".format( upload_bps=self.upload_bps, side_b_pdu=self.side_b_pdu, sta_list=sta_list) results_dict['numeric-score'] = "{}".format(udp_ul) results_dict['Units'] = "bps" self.kpi_csv.kpi_csv_write_dict(results_dict) # kpi data for TCP download traffic results_dict = self.kpi_csv.kpi_csv_get_dict_update_time() results_dict['Graph-Group'] = "TCP Download Rate" results_dict['pass/fail'] = pass_fail results_dict['Subtest-Pass'] = subpass_tcp_dl results_dict['Subtest-Fail'] = subfail_tcp_dl results_dict['short-description'] = "TCP-DL {download_bps} bps pdu {side_a_pdu} {sta_list} STA".format( download_bps=self.download_bps, side_a_pdu=self.side_a_pdu, sta_list=sta_list) results_dict['numeric-score'] = "{}".format(tcp_dl) results_dict['Units'] = "bps" self.kpi_csv.kpi_csv_write_dict(results_dict) # kpi data for TCP upload traffic results_dict['Graph-Group'] = "TCP Upload Rate" results_dict['pass/fail'] = pass_fail results_dict['Subtest-Pass'] = subpass_tcp_ul results_dict['Subtest-Fail'] = subfail_tcp_ul results_dict['short-description'] = "TCP-UL {upload_bps} bps pdu {side_b_pdu} {sta_list} STA".format( upload_bps=self.upload_bps, side_b_pdu=self.side_b_pdu, sta_list=sta_list) results_dict['numeric-score'] = "{}".format(tcp_ul) results_dict['Units'] = "bps" self.kpi_csv.kpi_csv_write_dict(results_dict) # record results for .html & .pdf reports def record_results( self, sta_count, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl_bps, total_ul_bps): dl = self.download_bps ul = self.upload_bps dl_pdu = self.side_a_pdu ul_pdu = self.side_b_pdu tags = dict() tags['requested-ul-bps'] = ul tags['requested-dl-bps'] = dl tags['ul-pdu-size'] = ul_pdu tags['dl-pdu-size'] = dl_pdu tags['station-count'] = sta_count # tags['attenuation'] = atten tags["script"] = 'sta_connect2' ''' # Add user specified tags for k in self.user_tags: tags[k[0]] = k[1] ''' # now = str(datetime.datetime.utcnow().isoformat()) print( "NOTE: Adding results to influx, total-download-bps: %s upload: %s bi-directional: %s\n" % (total_dl_bps, total_ul_bps, (total_ul_bps + total_dl_bps))) ''' if self.influxdb is not None: self.influxdb.post_to_influx( "total-download-bps", total_dl_bps, tags, now) self.influxdb.post_to_influx( "total-upload-bps", total_ul_bps, tags, now) self.influxdb.post_to_influx( "total-bi-directional-bps", total_ul_bps + total_dl_bps, tags, now) ''' if self.csv_results_file: row = [self.epoch_time, self.time_stamp(), sta_count, ul, ul, dl, dl, dl_pdu, dl_pdu, ul_pdu, ul_pdu, udp_ul, tcp_ul, udp_dl, tcp_dl, total_dl_bps, total_ul_bps, (total_ul_bps + total_dl_bps) ] ''' # Add values for any user specified tags for k in self.user_tags: row.append(k[1]) ''' self.csv_results_writer.writerow(row) self.csv_results_file.flush() def csv_generate_results_column_headers(self): csv_rx_headers = [ 'Time epoch', 'Time', 'Station-Count', 'UL-Min-Requested', 'UL-Max-Requested', 'DL-Min-Requested', 'DL-Max-Requested', 'UL-Min-PDU', 'UL-Max-PDU', 'DL-Min-PDU', 'DL-Max-PDU', # 'Attenuation', 'UDP-Upload-bps', 'TCP-Upload-bps', 'UDP-Download-bps', 'TCP-Download-bps', 'Total-UDP/TCP-Upload-bps', 'Total-UDP/TCP-Download-bps', 'Total-UDP/TCP-UL/DL-bps'] return csv_rx_headers # Write initial headers to csv file. def csv_add_column_headers(self): if self.csv_results_file is not None: self.csv_results_writer.writerow( self.csv_generate_results_column_headers()) self.csv_results_file.flush() # ~class # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - def main(): lfjson_host = "localhost" lfjson_port = 8080 parser = argparse.ArgumentParser( prog="sta_connect2.py", formatter_class=argparse.RawTextHelpFormatter, description=""" --------------------------- LANforge Unit Test: Connect Station to AP - sta_connect2.py --------------------------- Summary: This will create a station, create TCP and UDP traffic, run it a short amount of time, and verify whether traffic was sent and received. It also verifies the station connected to the requested BSSID if bssid is specified as an argument. The script will clean up the station and connections at the end of the test. --------------------------- CLI Example: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth1 --radio wiphy1 CLI Example for kpi.csv report output: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth2 --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234 CLI Example for kpi.csv, variable tx/rx rates, and pdu size: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth2 --download_bps 768000 --upload_bps 256000 --side_a_pdu 300 --side_b_pdu 750 --csv_outfile sta_connect2.csv --test_rig LF-Lab --test_tag L3 --dut_hw_version Linux --dut_model_num 1 --dut_sw_version 5.4.5 --dut_serial_num 1234 Note: --sta_mode use values in second column AUTO | 0 # 802.11g 802.11a | 1 # 802.11a b | 2 # 802.11b g | 3 # 802.11g abg | 4 # 802.11abg abgn | 5 # 802.11abgn bgn | 6 # 802.11bgn bg | 7 # 802.11bg abgnAC | 8 # 802.11abgn-AC anAC | 9 # 802.11an-AC an | 10 # 802.11an bgnAC | 11 # 802.11bgn-AC abgnAX | 12 # 802.11abgn-AX # a/b/g/n/AC/AX (dual-band AX) support bgnAX | 13 # 802.11bgn-AX anAX | 14 # 802.11an-AX aAX | 15 # 802.11a-AX (6E disables /n and /ac) --------------------------- """, epilog=""" --------------------------- CLI Example: ./sta_connect2.py --mgr localhost --dut_ssid --dut_passwd --dut_security wpa2 --upstream_port eth2 --radio wiphy1 """ ) parser.add_argument("-m", "--mgr", type=str, help="address of the LANforge GUI machine (localhost is default)", default='localhost') parser.add_argument("-o", "--port", type=int, help="IP Port the LANforge GUI is listening on (8080 is default)", default=8080) parser.add_argument("-u", "--user", type=str, help="TBD: credential login/username") parser.add_argument("-p", "--passwd", type=str, help="TBD: credential password") parser.add_argument("--resource", type=str, help="LANforge Station resource ID to use, default is 1", default=1) parser.add_argument("--upstream_resource", type=str, help="LANforge Ethernet port resource ID to use, default is 1", default=None) parser.add_argument("--upstream_port", type=str, help="LANforge Ethernet port name, default is eth2", default='1.1.eth2') parser.add_argument("--radio", type=str, help="LANforge radio to use, default is wiphy0", default='wiphy0') parser.add_argument("--sta_mode", type=str, help="LANforge station-mode setting (see add_sta LANforge CLI documentation, default is 0 (auto))", default=0) parser.add_argument("--dut_ssid", type=str, help="DUT SSID") parser.add_argument("--dut_security", type=str, help="DUT security: openLF, wpa, wpa2, wpa3") parser.add_argument("--dut_passwd", type=str, help="DUT PSK password. Do not set for OPEN auth") parser.add_argument("--dut_bssid", type=str, help="DUT BSSID to which we expect to connect.") parser.add_argument("--download_bps", type=int, help="Set the minimum bps value on test endpoint A. Default: 25g000", default=256000) parser.add_argument("--upload_bps", type=int, help="Set the minimum bps value on test endpoint B. Default: 256000", default=256000) parser.add_argument("--side_a_pdu", type=int, help="Set the minimum pdu value on test endpoint A. Default: 1200", default=1200) parser.add_argument("--side_b_pdu", type=int, help="Set the minimum pdu value on test endpoint B. Default: 1500", default=1500) parser.add_argument("--runtime_sec", type=int, help="Set test duration time. Default: 60 seconds", default=60) parser.add_argument("--debug", help="enable debugging", action="store_true") parser.add_argument("--prefix", type=str, help="Station prefix. Default: 'sta'", default='sta') parser.add_argument("--bringup_time", type=int, help="Seconds to wait for stations to associate and aquire IP. Default: 300", default=300) parser.add_argument('--influx_user', help='Username for your Influx database', default=None) parser.add_argument('--influx_passwd', help='Password for your Influx database', default=None) parser.add_argument('--influx_db', help='Name of your Influx database', default=None) parser.add_argument('--influx_host', help='Host of your influx database if different from the system you are running on', default='localhost') parser.add_argument('--monitor_interval', help='How frequently you want to append to your database', default='5s') parser.add_argument('--debug_log', default=None, help="Specify a file to send debug output to") parser.add_argument('--no_cleanup', help='Do not cleanup before exit', action='store_true') parser.add_argument('--local_lf_report_dir', help='--local_lf_report_dir override the report path, primary use when running test in test suite', default="") # kpi_csv arguments: parser.add_argument( "--test_rig", default="", help="test rig for kpi.csv, testbed that the tests are run on") parser.add_argument( "--test_tag", default="", help="test tag for kpi.csv, test specific information to differenciate the test") parser.add_argument( "--dut_hw_version", default="", help="dut hw version for kpi.csv, hardware version of the device under test") parser.add_argument( "--dut_sw_version", default="", help="dut sw version for kpi.csv, software version of the device under test") parser.add_argument( "--dut_model_num", default="", help="dut model for kpi.csv, model number / name of the device under test") parser.add_argument( "--dut_serial_num", default="", help="dut serial for kpi.csv, serial number / serial number of the device under test") parser.add_argument( "--test_priority", default="", help="dut model for kpi.csv, test-priority is arbitrary number") parser.add_argument( "--test_id", default="Layer 3 Traffic", help="test-id for kpi.csv, script or test name") parser.add_argument( '--csv_outfile', help="--csv_outfile ", default="") # logging configuration: parser.add_argument('--log_level', default=None, help='Set logging level: debug | info | warning | error | critical') parser.add_argument("--lf_logger_config_json", help="--lf_logger_config_json , json configuration of logger") args = parser.parse_args() # set up logger logger_config = lf_logger_config.lf_logger_config() # set the logger level to debug if args.log_level: logger_config.set_level(level=args.log_level) # lf_logger_config_json will take presidence to changing debug levels if args.lf_logger_config_json: # logger_config.lf_logger_config_json = "lf_logger_config.json" logger_config.lf_logger_config_json = args.lf_logger_config_json logger_config.load_lf_logger_config() # for kpi.csv generation local_lf_report_dir = args.local_lf_report_dir test_rig = args.test_rig test_tag = args.test_tag dut_hw_version = args.dut_hw_version dut_sw_version = args.dut_sw_version dut_model_num = args.dut_model_num dut_serial_num = args.dut_serial_num # test_priority = args.test_priority # this may need to be set per test test_id = args.test_id if local_lf_report_dir != "": report = lf_report.lf_report( _path=local_lf_report_dir, _results_dir_name="sta_connect2", _output_html="sta_connect2.html", _output_pdf="sta_connect2.pdf") else: report = lf_report.lf_report( _results_dir_name="sta_connect2", _output_html="sta_connect2.html", _output_pdf="sta_connect2.pdf") kpi_path = report.get_report_path() # kpi_filename = "kpi.csv" logger.info("kpi_path :{kpi_path}".format(kpi_path=kpi_path)) kpi_csv = lf_kpi_csv.lf_kpi_csv( _kpi_path=kpi_path, _kpi_test_rig=test_rig, _kpi_test_tag=test_tag, _kpi_dut_hw_version=dut_hw_version, _kpi_dut_sw_version=dut_sw_version, _kpi_dut_model_num=dut_model_num, _kpi_dut_serial_num=dut_serial_num, _kpi_test_id=test_id) if args.csv_outfile is not None: current_time = time.strftime("%Y-%m-%d-%H-%M-%S", time.localtime()) csv_outfile = "{}_{}-sta_connect2.csv".format( args.csv_outfile, current_time) csv_outfile = report.file_add_path(csv_outfile) logger.info("csv output file : {}".format(csv_outfile)) upstream_port = LFUtils.name_to_eid(args.upstream_port) logger.info("upstream_port: %s", upstream_port) if args.upstream_resource: upstream_resource = args.upstream_resource else: upstream_resource = upstream_port[1] staConnect = StaConnect2(args.mgr, args.port, _resource=args.resource, _upstream_resource=upstream_resource, _upstream_port=upstream_port[2], _radio=args.radio, _sta_mode=args.sta_mode, debug_=args.debug, _influx_db=args.influx_db, _influx_passwd=args.influx_passwd, _influx_user=args.influx_user, _influx_host=args.influx_host, kpi_csv=kpi_csv, outfile=args.csv_outfile, download_bps=args.download_bps, upload_bps=args.upload_bps, side_a_pdu=args.side_a_pdu, side_b_pdu=args.side_b_pdu, _runtime_sec=args.runtime_sec, _exit_on_fail=True, _exit_on_error=False) if args.user: staConnect.user = args.user if args.passwd: staConnect.passwd = args.passwd if args.dut_ssid: staConnect.dut_ssid = args.dut_ssid if args.dut_passwd: staConnect.dut_passwd = args.dut_passwd if args.dut_bssid: staConnect.dut_bssid = args.dut_bssid if args.dut_security: staConnect.dut_security = args.dut_security if args.prefix or (args.prefix != "sta"): args.prefix = staConnect.sta_prefix staConnect.station_names = ["%s0000" % args.prefix] staConnect.bringup_time_sec = args.bringup_time staConnect.setup() staConnect.start() logger.info("running traffic for %f sec", staConnect.runtime_secs) time.sleep(staConnect.runtime_secs) # fill out data kpi.csv and results reports temp_stations_list = [] temp_stations_list.extend(staConnect.station_profile.station_names.copy()) total_test = len(staConnect.get_result_list()) total_pass = len(staConnect.get_passed_result_list()) endp_rx_map, endp_rx_drop_map, endps, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul, total_dl_ll, total_ul_ll = staConnect.get_rx_values() staConnect.record_kpi_csv(len(temp_stations_list), total_test, total_pass, udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul) staConnect.record_results(len(temp_stations_list), udp_dl, tcp_dl, udp_ul, tcp_ul, total_dl, total_ul) # Reporting Results (.pdf & .html) csv_results_file = staConnect.get_csv_name() logger.info("csv_results_file: {}".format(csv_results_file)) # csv_results_file = kpi_path + "/" + kpi_filename report.set_title("Verify Layer 3 Throughput Over Time: sta_connect2.py") report.build_banner_left() report.start_content_div2() report.set_obj_html("Objective", "The Station Connection 2 Test is designed to test the performance of the " "Access Point with running TCP and UDP traffic for an amount of time. " "Verifies the station connected to the requested BSSID if bssid is specified." ) report.build_objective() test_setup_info = { "DUT Name": args.dut_model_num, "DUT Hardware Version": args.dut_hw_version, "DUT Software Version": args.dut_sw_version, "DUT Serial Number": args.dut_serial_num, "SSID": args.dut_ssid, } report.set_table_title("Device Under Test Information") report.build_table_title() report.test_setup_table(value="Device Under Test", test_setup_data=test_setup_info) test_input_info = { "LANforge ip": args.mgr, "LANforge port": args.port, "LANforge resource": args.resource, "Upstream": args.upstream_port, "Radio": args.radio, "SSID": args.dut_ssid, "Security": args.dut_security, "A side pdu": args.side_a_pdu, "B side pdu": args.side_b_pdu, "Download bps": args.download_bps, "Upload bps": args.upload_bps, "Test Duration": args.runtime_sec, } report.set_table_title("Test Configuration") report.build_table_title() report.test_setup_table(value="Test Configuration", test_setup_data=test_input_info) report.set_table_title("Layer 3 Station Connect 2 Traffic Results") report.build_table_title() report.set_table_dataframe_from_csv(csv_results_file) report.build_table() report.build_footer() report.write_html_with_timestamp() report.write_index_html() # report.write_pdf(_page_size = 'A3', _orientation='Landscape') # report.write_pdf_with_timestamp(_page_size='A4', _orientation='Portrait') # TODO install wkhtmltopdf on windows # if platform.system() == 'Linux': # report.write_pdf_with_timestamp(_page_size='A4', _orientation='Landscape') if platform.system() == 'Linux': report.write_pdf_with_timestamp(_page_size='letter', _orientation='portrait') staConnect.stop() # exit(1) # py-json/lanforge/lfcli_base.py - get_result_list(): staConnect.get_result_list() is_passing = staConnect.passes() # py-json/lanforge/lfcli_base.py - get_all_message(): logger.info(staConnect.get_all_message()) # cleanup stations if not args.no_cleanup: staConnect.cleanup() # Print path to results logger.info("Test Results HTML located: {report}".format(report=report.write_output_html)) if platform.system() == 'Linux': logger.info("Test Results PDF located: {report}".format(report=report.write_output_pdf)) # Added Exit codes if not is_passing: logger.info("FAIL: Some tests failed") exit(1) else: logger.info("PASS: All tests pass") exit(0) # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - if __name__ == "__main__": main()