JohnnyBoy00 commited on
Commit
47d38a9
1 Parent(s): 150c826

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -2,7 +2,6 @@
2
  language: en
3
  datasets:
4
  - JohnnyBoy00/saf_communication_networks_english
5
- license: apache-2.0
6
  tags:
7
  - generated_from_trainer
8
  widget:
@@ -67,9 +66,8 @@ The following hyperparameters were used during training:
67
  - train_batch_size: 1
68
  - gradient_accumulation_steps: 4
69
  - eval_batch_size: 4
70
- - seed: 42
71
  - mixed_precision_training: Native AMP
72
- - total_train_batch_size: 4
73
 
74
  ### Framework versions
75
 
@@ -100,7 +98,7 @@ The example below shows how the model can be applied to generate feedback to a g
100
  from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
101
  model = AutoModelForSeq2SeqLM.from_pretrained('JohnnyBoy00/bart-finetuned-saf-communication-networks')
102
  tokenizer = AutoTokenizer.from_pretrained('JohnnyBoy00/bart-finetuned-saf-communication-networks')
103
- example_input = ''
104
  inputs = tokenizer(example_input, max_length=256, padding='max_length', truncation=True, return_tensors='pt')
105
  generated_tokens = model.generate(
106
  inputs['input_ids'],
@@ -113,7 +111,7 @@ output = tokenizer.batch_decode(generated_tokens, skip_special_tokens=True)[0]
113
  The output produced by the model then looks as follows:
114
 
115
  ```
116
- Correct Feedback:
117
  ```
118
 
119
  ## Related Work
 
2
  language: en
3
  datasets:
4
  - JohnnyBoy00/saf_communication_networks_english
 
5
  tags:
6
  - generated_from_trainer
7
  widget:
 
66
  - train_batch_size: 1
67
  - gradient_accumulation_steps: 4
68
  - eval_batch_size: 4
 
69
  - mixed_precision_training: Native AMP
70
+ - seed: 42
71
 
72
  ### Framework versions
73
 
 
98
  from transformers import AutoModelForSeq2SeqLM, AutoTokenizer
99
  model = AutoModelForSeq2SeqLM.from_pretrained('JohnnyBoy00/bart-finetuned-saf-communication-networks')
100
  tokenizer = AutoTokenizer.from_pretrained('JohnnyBoy00/bart-finetuned-saf-communication-networks')
101
+ example_input = 'Answer: In TCP there is a Sequence Number field to identify packets individually for reliability. There is no Sequence Number in UDP. The UDP header does not have an options field, while the TCP header does. In TCP there is an Advertised Window field for the Sliding Window Protocol for Flow Control. There is no Flow Control and therefore no Advertised Window field in UDP. In TCP there there is only a Data Offset field that specifies the header length. In UDP the whole Packet Length is transmitted. Reference: Possible Differences : The UPD header (8 bytes) is much shorter than the TCP header (20-60 bytes) The UDP header has a fixed length while the TCP header has a variable length Fields contained in the TCP header and not the UDP header : -Sequence number -Acknowledgment number -Reserved -Flags/Control bits -Advertised window -Urgent Pointer -Options + Padding if the options are UDP includes the packet length (data + header) while TCP has the header length/data offset (just header) field instead The sender port field is optional in UDP, while the source port in TCP is necessary to establish the connection Question: State at least 4 of the differences shown in the lecture between the UDP and TCP headers.'
102
  inputs = tokenizer(example_input, max_length=256, padding='max_length', truncation=True, return_tensors='pt')
103
  generated_tokens = model.generate(
104
  inputs['input_ids'],
 
111
  The output produced by the model then looks as follows:
112
 
113
  ```
114
+ Correct Feedback: The response correctly identifies four differences between TCP and UDP headers.
115
  ```
116
 
117
  ## Related Work