In Section~\ref{sec:thething}, we introduced the notion of {\thething} events in privacy-preserving time series publishing.
The differentiation among regular and {\thething} events stipulates a privacy budget allocation that deviates from the application of existing differential privacy protection levels.
For the latter, a potentially adversarial data analyst may infer $L$ by observing the values of the privacy budget, which is usually an inseparable attribute of the data release as an indicator of the privacy guarantee to the users and as an estimate of the data utility to the analysts.
Hence, in both cases, a user-defined $L$, which is supposed to facilitate the configurable privacy protection of the user, could end up posing a privacy risk to them.
In Example~\ref{ex:lmdk-risk}, we demonstrate the extreme case of the application of the \texttt{Skip}{\thething} privacy scheme from Figure~\ref{fig:lmdk-skip}, where we approximate {\thethings} with the latest data release and invest all of the available privacy budget to regular events.
Figure~\ref{fig:lmdk-risk} shows the privacy risk that the application of a {\thething} privacy scheme that nullifies or approximates outputs, similar to \texttt{Skip}, might cause.
We point out in red the details that might cause indirect information inference.
In this extreme case, the minimization of the privacy budget in combination with nullifying the output (either by not publishing or by adding a lot of noise) or approximating the current output with previously released outputs might hint to any adversary that the current event is a {\thething}.
Therefore, an adversary who observes the values of the privacy budget can easily infer not only the number but also the exact temporal position of the {\thethings}.