Due to a ransomware attack, the wiki was reverted to a July 2022 version. . We apologize for the lack of a more recent valid backup.
...
| HTML |
|---|
<img src="https://jenkins.onosproject.org/view/QA/job/postjob-VM/lastSuccessfulBuild/artifact/HAfullNetPartition_master_20-builds_graph.jpg", alt="HAfullNetPartition", style="width:525px;height:350px;border:0"> |
commit 5de46cdd4fc9ac67f0cd800a4179b13b4c45e043 5c2807cf9dd89c7cb2222f6fffc0047e21fd4a49 (HEAD, origin/master, origin/HEAD, master)
Author: Lakshya Thakur Carmelo Cascone [lapstjup@gmailcarmelo@opennetworking.comorg]
AuthorDate: Fri Mon Aug 14 193 01:3212:07 10 2020 +0530-0700
Commit: Lakshya Thakur [lapstjup@gmail.comPier Luigi Ventre [pier@opennetworking.org]
CommitDate: Fri Aug 14 Tue Sep 1 19:3253:07 16 2020 +0530
Added io.atomix prefix to enable TLS to support single node and multi node cluster setup0000
Use bazelisk inside Dockerfile
--
(cherry picked from commit 5524134a4ea7914d5e814e7f4acf496a68044790)
Case 1: Constructing test variables and building ONOS package - PASS
...
- 17.1 Increment then get a default counter on each node - PASS
- 17.2 Get then Increment a default counter on each node - PASS
- 17.3 Counters we added have the correct values - PASS
- 17.4 Add -8 to then get a default counter on each node - PASS
- 17.5 Add 5 to then get a default counter on each node - PASS
- 17.6 Get then add 5 to a default counter on each node - PASS
- 17.7 Counters we added have the correct values - PASS
- 17.8 Distributed Set get - PASS
- 17.9 Distributed Set size - PASS
- 17.10 Distributed Set add() - PASS
- 17.11 Distributed Set addAll() - PASS
- 17.12 Distributed Set contains() - PASS
- 17.13 Distributed Set containsAll() - PASS
- 17.14 Distributed Set remove() - PASS
- 17.15 Distributed Set removeAll() - PASS
- 17.16 Distributed Set addAll() - PASS
- 17.17 Distributed Set clear() - PASS
- 17.18 Distributed Set addAll() - PASS
- 17.19 Distributed Set retain() - FAIL
- Set retain was incorrect
- 17.20 Partitioned Transactional maps put - PASS
- 17.21 Partitioned Transactional maps get - PASS
- 17.22 Get the value of a new value - PASS
- 17.23 Atomic Value set() - PASS
- 17.24 Get the value after set() - PASS
- 17.25 Atomic Value compareAndSet() - PASS
- 17.26 Get the value after compareAndSet() - PASS
- 17.27 Atomic Value getAndSet() - PASS
- 17.28 Get the value after getAndSet() - PASS
- 17.29 Atomic Value destory() - PASS
- 17.30 Get the value after destroy() - PASS
- 17.31 Work Queue add() - PASS
- 17.32 Check the work queue stats - PASS
- 17.33 Work Queue addMultiple() - PASS
- 17.34 Check the work queue stats - PASS
- 17.35 Work Queue takeAndComplete() 1 - PASS
- 17.36 Check the work queue stats - PASS
- 17.37 Work Queue takeAndComplete() 2 - PASS
- 17.38 Check the work queue stats - PASS
- 17.39 Work Queue destroy() - PASS
- 17.40 Check the work queue stats - PASS
Case 61: Partition ONOS nodes into two distinct partitions - PASS
- 61.1 Checking ONOS Logs for errors - No Result
- 61.2 Partitioning ONOS nodes - PASS
- 61.3 Sleeping 60 seconds - No Result
Case 8: Compare ONOS Topology view to Mininet topology - PASS
Compare topology objects between Mininet and ONOS
- 8.1 Comparing ONOS topology to MN topology - PASS
- 8.2 Hosts view is consistent across all ONOS nodes - PASS
- 8.3 Hosts information is correct - PASS
- 8.4 Host attachment points to the network - PASS
- 8.5 Clusters view is consistent across all ONOS nodes - PASS
- 8.6 There is only one SCC - PASS
- 8.7 Device information is correct - PASS
- 8.8 Links are correct - PASS
- 8.9 Hosts are correct - PASS
- 8.10 Checking ONOS nodes - PASS
Case 7: Running ONOS Constant State Tests - FAIL
- 7.1 Check that each switch has a master - PASS
- 7.2 Read device roles from ONOS - PASS
- 7.3 Check for consistency in roles from each controller - PASS
- 7.4 Get the intents from each controller - PASS
- 7.5 Check for consistency in Intents from each controller - PASS
- 7.6 Compare current intents with intents before the failure - PASS
- 7.7 Get the OF Table entries and compare to before component failure - FAIL
- Changes were found in the flow tables
- 7.8 Leadership Election is still functional - PASS
Case 4: Verify connectivity by sending traffic across Intents - FAIL
Ping across added host intents to check functionality and check the state of the intent
- 4.1 Check Intent state - PASS
- 4.2 Ping across added host intents - FAIL
- Intents have not been installed correctly, pings failed.
- 4.3 Check leadership of topics - PASS
- 4.4 Wait a minute then ping again - FAIL
- Intents have not been installed correctly, pings failed.
Case 15: Check that Leadership Election is still functional - PASS
- 15.1 Run for election on each node - PASS
- 15.2 Check that each node shows the same leader and candidates - PASS
- 15.3 Find current leader and withdraw - PASS
- 15.4 Check that a new node was elected leader - PASS
- 15.5 Check that that new leader was the candidate of old leader - PASS
- 15.6 Run for election on old leader( just so everyone is in the hat ) - PASS
- 15.7 Check that oldLeader is a candidate, and leader if only 1 node - No Result
Case 17: Check for basic functionality with distributed primitives - FAIL
Test the methods of the distributed primitives (counters and sets) throught the cli
- 17.1 Increment then get a default counter on each node - PASS
- 17.2 Get then Increment a default counter on each node - PASS
- 17.3 Counters we added have the correct values - PASS
- 17.4 Add -8 to then get a default counter on each node - PASS
- 17.5 Add 5 to then get a default counter on each node - PASS
- 17.6 Get then add 5 to a default counter on each node - PASS
- 17.7 Counters we added have the correct values - PASS
- 17.8 Distributed Set get - FAIL
- Set elements are incorrect
- 17.9 Distributed Set size - FAIL
- Set sizes are incorrect
- 17.10 Distributed Set add() - PASS
- 17.11 Distributed Set addAll() - PASS
- 17.12 Distributed Set contains() - PASS
- 17.13 Distributed Set containsAll() - PASS
- 17.14 Distributed Set remove() - PASS
- 17.15 Distributed Set removeAll() - PASS
- 17.16 Distributed Set addAll() - PASS
- 17.17 Distributed Set clear() - PASS
- 17.18 Distributed Set addAll() - PASS
- 17.19 Distributed Set retain() - FAIL
- Set retain was incorrect
- 17.20 Partitioned Transactional maps put - PASS
- 17.21 Partitioned Transactional maps get - PASS
- 17.22 Get the value of a new value - PASS
- 17.23 Atomic Value set() - PASS
- 17.24 Get the value after set() - PASS
- 17.25 Atomic Value compareAndSet() - PASS
- 17.26 Get the value after compareAndSet() - PASS
- 17.27 Atomic Value getAndSet() - PASS
- 17.28 Get the value after getAndSet() - PASS
- 17.29 Atomic Value destory() - PASS
- 17.30 Get the value after destroy() - PASS
- 17.31 Work Queue add() - PASS
- 17.32 Check the work queue stats - PASS
- 17.33 Work Queue addMultiple() - PASS
- 17.34 Check the work queue stats - PASS
- 17.35 Work Queue takeAndComplete() 1 - PASS
- 17.36 Check the work queue stats - PASS
- 17.37 Work Queue takeAndComplete() 2 - PASS
- 17.38 Check the work queue stats - PASS
- 17.39 Work Queue destroy() - PASS
- 17.40 Check the work queue stats - PASS
Case 62: Healing Partition - PASS
- 62.1 Deleteing firewall rules - PASS
- 62.2 Checking ONOS nodes - PASS
Case 8: Compare ONOS Topology view to Mininet topology - FAIL
Compare topology objects between Mininet and ONOS
- 8.1 Comparing ONOS topology to MN topology - FAIL
- ONOS topology don't match Mininet
- 8.2 Hosts view is consistent across all ONOS nodes - PASS
- 8.3 Hosts information is correct - PASS
- 8.4 Host attachment points to the network - PASS
- 8.5 Clusters view is consistent across all ONOS nodes - FAIL
- ONOS nodes have different views of clusters
- 8.6 There is only one SCC - FAIL
- ONOS shows 22 SCCs
- 8.7 Device information is correct - PASS
- 8.8 Links are correct - FAIL
- Links are incorrect
- 8.9 Hosts are correct - PASS
- 8.10 Checking ONOS nodes - PASS
Case 7: Running ONOS Constant State Tests - FAIL
- 7.1 Check that each switch has a master - PASS
- 7.2 Read device roles from ONOS - PASS
- 7.3 Check for consistency in roles from each controller - FAIL
- ONOS nodes have different views of switch roles
- 7.4 Get the intents from each controller - PASS
- 7.5 Check for consistency in Intents from each controller - PASS
- 7.6 Compare current intents with intents before the failure - FAIL
- The Intents changed during failure
- 7.7 Get the OF Table entries and compare to before component failure - FAIL
- Changes were found in the flow tables
- 7.8 Leadership Election is still functional - FAIL
- Something went wrong with Leadership election
Case 4: Verify connectivity by sending traffic across Intents - FAIL
...
- 4.1 Check Intent state - FAIL
- Intents are not all in INSTALLED state
- 4.2 Ping across added host intents - FAIL
- Intents have not been installed correctly, pings failed.
- 4.3 Check leadership of topics - PASS
- 4.4 Wait a minute then ping again - FAIL
- Intents have not been installed correctly, pings failed.
Case 15: Check that Leadership Election is still functional - PASS
- 15.1 Run for election on each node - PASS
- 15.2 Check that each node shows the same leader and candidates - PASS
- 15.3 Find current leader and withdraw - PASS
- 15.4 Check that a new node was elected leader - PASS
- 15.5 Check that that new leader was the candidate of old leader - PASS
- 15.6 Run for election on old leader( just so everyone is in the hat ) - PASS
- 15.7 Check that oldLeader is a candidate, and leader if only 1 node - No Result
Case 17: Check for basic functionality with distributed primitives - FAIL
Test the methods of the distributed primitives (counters and sets) throught the cli
- 17.1 Increment then get a default counter on each node - PASS
- 17.2 Get then Increment a default counter on each node - PASS
- 17.3 Counters we added have the correct values - PASS
- 17.4 Add -8 to then get a default counter on each node - PASS
- 17.5 Add 5 to then get a default counter on each node - PASS
- 17.6 Get then add 5 to a default counter on each node - PASS
- 17.7 Counters we added have the correct values - PASS
- 17.8 Distributed Set get - FAIL
- Set elements are incorrect
- 17.9 Distributed Set size - FAIL
- Set sizes are incorrect
- 17.10 Distributed Set add() - PASS
- 17.11 Distributed Set addAll() - PASS
- 17.12 Distributed Set contains() - PASS
- 17.13 Distributed Set containsAll() - PASS
- 17.14 Distributed Set remove() - PASS
- 17.15 Distributed Set removeAll() - PASS
- 17.16 Distributed Set addAll() - PASS
- 17.17 Distributed Set clear() - PASS
- 17.18 Distributed Set addAll() - PASS
- 17.19 Distributed Set retain() - FAIL
- Set retain was incorrect
- 17.20 Partitioned Transactional maps put - PASS
- 17.21 Partitioned Transactional maps get - FAIL
- Partitioned Transactional Map values incorrect
- 17.22 Get the value of a new value - FAIL
- Error getting atomic Value None, found: [None, 'null', 'null', None, 'null']
- 17.23 Atomic Value set() - FAIL
- Error setting atomic Value[-1, 1, 1, -1, 1]
- 17.24 Get the value after set() - FAIL
- Error getting atomic Value foo, found: [None, 'foo', 'foo', None, 'foo']
- 17.25 Atomic Value compareAndSet() - FAIL
- Error setting atomic Value:-1
- 17.26 Get the value after compareAndSet() - FAIL
- Error getting atomic Value bar, found: [None, 'bar', 'bar', None, 'bar']
- 17.27 Atomic Value getAndSet() - PASS
- 17.28 Get the value after getAndSet() - FAIL
- Error getting atomic Value: expected baz, found: [None, 'baz', 'baz', None, 'baz']
- 17.29 Atomic Value destory() - FAIL
- Error destroying atomic Value
- PASS
- 17.30 Get the value after destroy() - FAIL
- Error getting atomic Value None, found: [None'baz', 'null', 'null', None, 'null']
- 17.31 Work Queue add() - PASS
FAIL
- Error adding to Work Queue
- 17.32 Check the work queue stats - No Result
Case
...
61: Partition ONOS nodes into two distinct partitions - PASS
- 61.1 Checking ONOS Logs for errors - No Result
- 61.2 Partitioning ONOS nodes - PASS
- 61.3 Sleeping 60 seconds - No Result
9.1 Kill Link between s28 and s3 - PASS