aboutsummaryrefslogtreecommitdiffstats
path: root/Zotlabs/Daemon/Externals.php
blob: 167b0524f990b452058bc3b923cc14b396dc26ba (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
<?php /** @file */

namespace Zotlabs\Daemon;

require_once('include/zot.php');
require_once('include/identity.php');


class Externals {

	static public function run($argc,$argv){

		$total = 0;
		$attempts = 0;

		logger('externals: startup', LOGGER_DEBUG);

		// pull in some public posts


		while($total == 0 && $attempts < 3) {
			$arr = array('url' => '');
			call_hooks('externals_url_select',$arr);

			if($arr['url']) {
				$url = $arr['url'];
			} 
			else {
				$randfunc = db_getfunc('RAND');

				// fixme this query does not deal with directory realms. 

				$r = q("select site_url, site_pull from site where site_url != '%s' and site_flags != %d and site_type = %d and site_dead = 0 order by $randfunc limit 1",
					dbesc(z_root()),
					intval(DIRECTORY_MODE_STANDALONE),
					intval(SITE_TYPE_ZOT)
				);
				if($r)
					$url = $r[0]['site_url'];
			}

			$blacklisted = false;

			if(! check_siteallowed($url)) {
				logger('blacklisted site: ' . $url);
				$blacklisted = true;
			}

			$attempts ++;

			// make sure we can eventually break out if somebody blacklists all known sites

			if($blacklisted) {
				if($attempts > 20)
					break;
				$attempts --;
				continue;
			}

			if($url) {
				if($r[0]['site_pull'] !== NULL_DATE)
					$mindate = urlencode(datetime_convert('','',$r[0]['site_pull'] . ' - 1 day'));
				else {
					$days = get_config('externals','since_days');
					if($days === false)
						$days = 15;
					$mindate = urlencode(datetime_convert('','','now - ' . intval($days) . ' days'));
				}

				$feedurl = $url . '/zotfeed?f=&mindate=' . $mindate;

				logger('externals: pulling public content from ' . $feedurl, LOGGER_DEBUG);

				$x = z_fetch_url($feedurl);
				if(($x) && ($x['success'])) {

					q("update site set site_pull = '%s' where site_url = '%s'",
						dbesc(datetime_convert()),
						dbesc($url)
					);

					$j = json_decode($x['body'],true);
					if($j['success'] && $j['messages']) {
						$sys = get_sys_channel();
						foreach($j['messages'] as $message) {
							// on these posts, clear any route info. 
							$message['route'] = '';
							$results = process_delivery(array('hash' => 'undefined'), get_item_elements($message),
								array(array('hash' => $sys['xchan_hash'])), false, true);
							$total ++;
						}
						logger('externals: import_public_posts: ' . $total . ' messages imported', LOGGER_DEBUG);
					}
				}
			}
		}
	}
}