Mitigation of attack vectors from external provided links for generic extractor
See original GitHub issueChecklist
- I’m asking a question
- I’ve looked through the README and FAQ for similar questions
- I’ve searched the bugtracker for similar questions including closed ones
Question
-
Are there any mitigation possibilities in place for attacks such as Slowloris or a 3 party url that will indefinitely send data without a
content-length
? -
Does the downloader respect the
content-length
header and error out if the downloaded size is a mismatch? -
I have noticed that there is an option for
--max-filesize
but does that usecontent-length
or the actual downloaded bytes so far? -
Is there anything I could do to kill the process if the download speed is lower than a threshold? Or is there any option that would help me achieve this?
Issue Analytics
- State:
- Created 3 years ago
- Comments:5
Top Results From Across the Web
What is an Attack Vector? 16 Common Attack Vectors in 2022
An attack vector is a path or means by which an attacker can gain unauthorized access to a computer or network to deliver...
Read more >8 Common Types of Cyber Attack Vectors and How to Avoid ...
Learn the 8 common cyber attack vectors and the steps you can take to mitigate it.
Read more >Malicious Link - an overview | ScienceDirect Topics
This attack is known as a drive-by download attack in which target users are coerced to visit malicious domains through social engineering [6]....
Read more >Guidance for preventing, detecting, and hunting for ... - Microsoft
This blog reports our observations and analysis of attacks that take advantage of the Log4j 2 vulnerabilities. It also provides our ...
Read more >A Threat-Driven Approach to Cyber Security - Lockheed Martin
The threat actor(s) gain access to the assets via attack vectors and vulnerabilities present in the technology components that house or provide direct...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
I’ve re-checked question 2 and looks like it actually silently truncates data at
Content-Length
border if there are more data. This happens somewhere in python internals or deeper. Browsers seem to have similar behavior. So that the actual answer to 2 is “no”.Thank you for your time & answers.
I have come to the conclusion that my requirements are out of the scope of
youtube-dl
.We can write a custom downloader that aligns with our needs and use that when the extractor is
generic
, else letyoutube-dl
handle the heavy work.I’ll most likely leverage
--load-info-json
to cache the metadata and not send out duplicate requests after the extractor checks.