This file is indexed.

/usr/share/plowshare/modules/bayfiles.sh is in plowshare-modules 0~git20160124.8a8190d-1.

This file is owned by root:root, with mode 0o644.

The actual contents of the file can be viewed below.

  1
  2
  3
  4
  5
  6
  7
  8
  9
 10
 11
 12
 13
 14
 15
 16
 17
 18
 19
 20
 21
 22
 23
 24
 25
 26
 27
 28
 29
 30
 31
 32
 33
 34
 35
 36
 37
 38
 39
 40
 41
 42
 43
 44
 45
 46
 47
 48
 49
 50
 51
 52
 53
 54
 55
 56
 57
 58
 59
 60
 61
 62
 63
 64
 65
 66
 67
 68
 69
 70
 71
 72
 73
 74
 75
 76
 77
 78
 79
 80
 81
 82
 83
 84
 85
 86
 87
 88
 89
 90
 91
 92
 93
 94
 95
 96
 97
 98
 99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
# Plowshare bayfiles.net module
# Copyright (c) 2012-2013 Plowshare team
#
# This file is part of Plowshare.
#
# Plowshare is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Plowshare is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Plowshare.  If not, see <http://www.gnu.org/licenses/>.

MODULE_BAYFILES_REGEXP_URL='https\?://\(www\.\)\?bayfiles\.net/'

MODULE_BAYFILES_DOWNLOAD_OPTIONS="
AUTH,a,auth,a=USER:PASSWORD,User account"
MODULE_BAYFILES_DOWNLOAD_RESUME=yes
MODULE_BAYFILES_DOWNLOAD_FINAL_LINK_NEEDS_COOKIE=no
MODULE_BAYFILES_DOWNLOAD_SUCCESSIVE_INTERVAL=300

MODULE_BAYFILES_UPLOAD_OPTIONS="
AUTH,a,auth,a=USER:PASSWORD,User account"
MODULE_BAYFILES_UPLOAD_REMOTE_SUPPORT=no

MODULE_BAYFILES_DELETE_OPTIONS=""
MODULE_BAYFILES_PROBE_OPTIONS=""

# Static function. Proceed with login (free or premium)
# Uses official API: http://bayfiles.net/api
bayfiles_login() {
    local AUTH=$1
    local API_URL=$2
    local USER PASSWORD LOGIN_JSON_DATA SESSID ERR

    split_auth "$AUTH" USER PASSWORD || return

    LOGIN_JSON_DATA=$(curl "${API_URL}/account/login/${USER}/${PASSWORD}") || return

    # {"error":"","session":"947qfkvd0eqvohb1sif3hcl0d2"}]
    SESSID=$(echo "$LOGIN_JSON_DATA" | parse_json_quiet 'session')
    if [ -z "$SESSID" ]; then
        ERR=$(echo "$LOGIN_JSON_DATA" | parse_json 'error')
        log_debug "Remote error: $ERR"
        return $ERR_LOGIN_FAILED
    fi

    log_debug "sessid: $SESSID"
    echo "$SESSID"
    return 0
}

# Output a bayfiles.net file download URL
# $1: cookie file (for account only)
# $2: bayfiles url
# stdout: real file download link
bayfiles_download() {
    local COOKIE_FILE=$1
    local URL=$2
    local API_URL='http://api.bayfiles.net/v1'
    local AJAX_URL='http://bayfiles.net/ajax_download'
    local PAGE FILE_URL FILENAME SESSION OPT_SESSION

    if [ -n "$AUTH" ]; then
        SESSION=$(bayfiles_login "$AUTH" "$API_URL") || return
        OPT_SESSION="-b SESSID=$SESSION"
        PAGE=$(curl -c "$COOKIE_FILE" $OPT_SESSION "$URL") || return
    else
        PAGE=$(curl -c "$COOKIE_FILE" -b "$COOKIE_FILE" "$URL") || return
    fi

    if match 'The link is incorrect\|<title>404 - Not Found</title>' "$PAGE"; then
        return $ERR_LINK_DEAD
    fi

    # <h3 class="comparison">What are the benefits for <strong>premium</strong> members?</h3>
    if match 'comparison\|benefits' "$PAGE"; then
        # Big files case
        local VFID DELAY TOKEN JSON_COUNT DATA_DL

        # Upgrade to premium or wait 5 minutes.
        if match 'Upgrade to premium or wait' "$PAGE"; then
            DELAY=$(echo "$PAGE" | parse 'premium or wait' \
                'wait[[:space:]]\([[:digit:]]\+\)[[:space:]]*minute')
            echo $((DELAY * 60))
            return $ERR_LINK_TEMP_UNAVAILABLE
        fi

        VFID=$(echo "$PAGE" | parse 'var vfid = ' '= \([[:digit:]]\+\);') || return

        # If no delay were found, we try without
        DELAY=$(echo "$PAGE" | parse_quiet 'var delay = ' '= \([[:digit:]]\+\);')

        JSON_COUNT=$(curl --get -b "$COOKIE_FILE" \
            --data "action=startTimer&vfid=$VFID" \
            "$AJAX_URL") || return

        TOKEN=$(echo "$JSON_COUNT" | parse_json token) || return

        wait $((DELAY)) || return

        DATA_DL=$(curl -b "$COOKIE_FILE" \
            $OPT_SESSION \
            --data "action=getLink&vfid=$VFID&token=$TOKEN" \
            "$AJAX_URL") || return

        FILE_URL=$(echo "$DATA_DL" | \
            parse 'onclick' "\(http[^']*\)") || return

    # Premium account
    else
        FILE_URL=$(echo "$PAGE" | parse_attr 'class="highlighted-btn' 'href') || return
        MODULE_BAYFILES_DOWNLOAD_SUCCESSIVE_INTERVAL=0
    fi

    # Extract filename from $PAGE, work for both cases
    FILENAME=$(parse_attr 'title="' 'title' <<< "$PAGE" | html_to_utf8)

    echo "$FILE_URL"
    echo "$FILENAME"
}

# Upload a file to bayfiles.net
# $1: cookie file (unused here)
# $2: input file (with full path)
# $3: remote filename
# stdout: download link + delete link + admin link
bayfiles_upload() {
    local FILE=$2
    local DESTFILE=$3
    local API_URL='http://api.bayfiles.net/v1'
    local SESSION_GET JSON UPLOAD_URL FILE_URL DELETE_URL ADMIN_URL

    # Account users (free or premium) have a session id
    if [ -n "$AUTH" ]; then
        SESSION_GET='?session='$(bayfiles_login "$AUTH" "$API_URL") || return
    else
        SESSION_GET=''
    fi

    JSON=$(curl "${API_URL}/file/uploadUrl${SESSION_GET}") || return

    # {"error":"","uploadUrl":"http ..","progressUrl":"http .."}
    UPLOAD_URL=$(echo "$JSON" | parse_json 'uploadUrl') || return

    # Sanity check (wrong upload url: site is broken)
    if [[ $UPLOAD_URL = http:///* ]]; then
        return $ERR_LINK_TEMP_UNAVAILABLE
    fi

    JSON=$(curl_with_log -F "file=@$FILE;filename=$DESTFILE" \
        "$UPLOAD_URL") || return

    # {"error":"","fileId":"abK1","size":"123456","sha1":"6f ..", ..}
    FILE_URL=$(echo "$JSON" | parse_json 'downloadUrl') || return
    DELETE_URL=$(echo "$JSON" | parse_json 'deleteUrl') || return
    ADMIN_URL=$(echo "$JSON" | parse_json 'linksUrl') || return

    echo "$FILE_URL"
    echo "$DELETE_URL"
    echo "$ADMIN_URL"
}

# Delete a file on bayfiles
# $1: cookie file (unused here)
# $2: delete link
bayfiles_delete() {
    local URL=$2
    local PAGE CONFIRM

    PAGE=$(curl "$URL") || return

    # Are you sure you want to <strong>delete</strong> this file?
    if match 'Confirm Deletion' "$PAGE"; then
        CONFIRM=$(echo "$PAGE" | parse_attr 'Confirm' href) || return
        PAGE=$(curl "$URL$CONFIRM") || return

        # File successfully deleted.
        match 'successfully deleted' "$PAGE" && return 0

    # The requested file could not be found.
    elif match 'file could not be found' "$PAGE"; then
        return $ERR_LINK_DEAD

    fi

    # Invalid security token. Please check your link.
    return $ERR_FATAL
}

# Probe a download URL
# $1: cookie file (unused here)
# $2: bayfile url
# $3: requested capability list
# stdout: 1 capability per line
bayfiles_probe() {
    local -r URL=$2
    local -r REQ_IN=$3
    local PAGE REQ_OUT FILE_SIZE

    PAGE=$(curl -L "$URL") || return

    if match 'The link is incorrect\|<title>404 - Not Found</title>' "$PAGE"; then
        return $ERR_LINK_DEAD
    fi

    REQ_OUT=c

    if [[ $REQ_IN = *f* ]]; then
        parse_attr 'title=' 'title' <<< "$PAGE" | html_to_utf8 && \
            REQ_OUT="${REQ_OUT}f"
    fi

    if [[ $REQ_IN = *s* ]]; then
        FILE_SIZE=$(echo "$PAGE" | parse '>File:<' '<strong>\([^<]*\)' 1) && \
            translate_size "$FILE_SIZE" && REQ_OUT="${REQ_OUT}s"
    fi

    echo $REQ_OUT
}